file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
combat.rs
|
pub fn target_tile(tcod: &mut Tcod,
objects: &[Object], game: &mut Game,
max_range: Option<f32>)
-> Option<(i32, i32)> {
use tcod::input::KeyCode::Escape;
loop {
// render the screen. this erases the inventory and shows the names of
// objects under the mouse.
tcod.root.flush();
let event = input::check_for_event(input::KEY_PRESS | input::MOUSE).map(|e| e.1);
let mut key = None;
match event {
Some(Event::Mouse(m)) => tcod.mouse = m,
Some(Event::Key(k)) => key = Some(k),
None => {}
}
render_all(tcod, objects, game, false);
let (x, y) = (tcod.mouse.cx as i32, tcod.mouse.cy as i32);
// accept the target if the player clicked in FOV, and in case a range
// is specified, if it's in that range
let in_fov = (x < MAP_WIDTH) && (y < MAP_HEIGHT) && tcod.fov.is_in_fov(x, y);
let in_range = max_range.map_or(
true, |range| objects[PLAYER].distance(x, y) <= range);
if tcod.mouse.lbutton_pressed && in_fov && in_range {
return Some((x, y))
}
let escape = key.map_or(false, |k| k.code == Escape);
if tcod.mouse.rbutton_pressed || escape {
return None // cancel if the player right-clicked or pressed Escape
}
}
}
/// find closest enemy, up to a maximum range, and in the player's FOV
pub fn closest_monster(max_range: i32, objects: &mut [Object], tcod: &Tcod) -> Option<usize> {
let mut closest_enemy = None;
let mut closest_dist = (max_range + 1) as f32; // start with (slightly more than) maximum range
for (id, object) in objects.iter().enumerate() {
if (id != PLAYER) && object.fighter.is_some() && object.ai.is_some() &&
tcod.fov.is_in_fov(object.x, object.y)
{
// calculate distance between this object and the player
let dist = objects[PLAYER].distance_to(object);
if dist < closest_dist { // it's closer, so remember it
closest_enemy = Some(id);
closest_dist = dist;
}
}
}
closest_enemy
}
pub fn cast_fireball(_inventory_id: usize, objects: &mut [Object],game: &mut Game, tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target tile to throw a fireball at
game.log.add("Left-click a target tile for the fireball, or right-click to cancel.",
colors::LIGHT_CYAN);
let (x, y) = match target_tile(tcod, objects, game, None) {
Some(tile_pos) => tile_pos,
None => return UseResult::Cancelled,
};
game.log.add(format!("The fireball explodes, burning everything within {} tiles!", FIREBALL_RADIUS),
colors::ORANGE);
let mut xp_to_gain = 0;
for (id, obj) in objects.iter_mut().enumerate() {
if obj.distance(x, y) <= FIREBALL_RADIUS as f32 && obj.fighter.is_some() {
game.log.add(format!("The {} gets burned for {} hit points.", obj.name, FIREBALL_DAMAGE),
colors::ORANGE);
if let Some(xp) = obj.take_damage(FIREBALL_DAMAGE, game) {
// Don't reward the player for burning themself!
if id != PLAYER {
xp_to_gain += xp;
}
}
}
}
objects[PLAYER].fighter.as_mut().unwrap().xp += xp_to_gain;
UseResult::UsedUp
}
pub fn cast_heal(_inventory_id: usize, objects: &mut [Object], game: &mut Game, _tcod: &mut Tcod)
-> UseResult
{
// heal the player
let player = &mut objects[PLAYER];
if let Some(fighter) = player.fighter {
if fighter.hp == player.max_hp(game) {
game.log.add("You are already at full health.", colors::RED);
return UseResult::Cancelled;
}
game.log.add("Your wounds start to feel better!", colors::LIGHT_VIOLET);
player.heal(HEAL_AMOUNT, game);
return UseResult::UsedUp;
}
UseResult::Cancelled
}
pub fn cast_lightning(_inventory_id: usize, objects: &mut [Object], game: &mut Game, tcod: &mut Tcod) -> UseResult
{
// find closest enemy (inside a maximum range) and damage it
let monster_id = closest_monster(LIGHTNING_RANGE, objects, tcod);
if let Some(monster_id) = monster_id {
// zap it!
game.log.add(format!("A lightning bolt strikes the {} with a loud thunder! \
The damage is {} hit points.",
objects[monster_id].name, LIGHTNING_DAMAGE),
colors::LIGHT_BLUE);
objects[monster_id].take_damage(LIGHTNING_DAMAGE, game);
UseResult::UsedUp
} else { // no enemy found within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn cast_confuse(_inventory_id: usize, objects: &mut [Object], game: &mut Game,tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target to confuse
game.log.add("Left-click an enemy to confuse it, or right-click to cancel.", colors::LIGHT_CYAN);
let monster_id = target_monster(tcod, objects, game, Some(CONFUSE_RANGE as f32));
if let Some(monster_id) = monster_id {
let old_ai = objects[monster_id].ai.take().unwrap_or(Ai::Basic);
// replace the monster's AI with a "confused" one; after
// some turns it will restore the old AI
objects[monster_id].ai = Some(Ai::Confused {
previous_ai: Box::new(old_ai),
num_turns: CONFUSE_NUM_TURNS,
});
game.log.add(format!("The eyes of {} look vacant, as he starts to stumble around!",
objects[monster_id].name),
colors::LIGHT_GREEN);
UseResult::UsedUp
} else { // no enemy fonud within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn player_death(player: &mut Object, messages: &mut Messages)
|
pub fn monster_death(monster: &mut Object, messages: &mut Messages) {
// transform it into a nasty corpse! it doesn't block, can't be
// attacked and doesn't move
// TODO Replace with game.log.add()
// message(messages, format!("{} is dead!", monster.name), colors::ORANGE);
message(messages, format!("{} is dead! You gain {} experience points.",
monster.name, monster.fighter.unwrap().xp), colors::ORANGE);
monster.char = CORPSE;
monster.color = colors::DARK_RED;
monster.blocks = false;
monster.fighter = None;
monster.ai = None;
monster.name = format!("remains of {}", monster.name);
}
pub fn player_move_or_attack(dx: i32, dy: i32, objects: &mut [Object], game: &mut Game) {
// the coordinates the player is moving to/attacking
let x = objects[PLAYER].x + dx;
let y = objects[PLAYER].y + dy;
// try to find an attackable object there
let target_id = objects.iter().position(|object| {
object.fighter.is_some() && object.pos() == (x, y)
});
// attack if target found, move otherwise
match target_id {
Some(target_id) => {
let (player, target) = mut_two(PLAYER, target_id, objects);
player.attack(target, game);
}
None => {
move_by(PLAYER, dx, dy, &mut game.map, objects);
}
}
}
pub fn ai_take_turn(monster_id: usize, objects: &mut [Object], game: &mut Game, fov_map: &FovMap) {
use Ai::*;
if let Some(ai) = objects[monster_id].ai.take() {
let new_ai = match ai {
Basic => ai_basic(monster_id, game, objects, fov_map),
Confused{previous_ai, num_turns} => ai_confused(
monster_id, &mut game.map, objects, &mut game.log, previous_ai, num_turns)
};
objects[monster_id].
|
{
// the game ended!
// TODO Replace with game.log.add()
message(messages, "You died!", colors::DARK_RED);
// for added effect, transform the player into a corpse!
player.char = CORPSE;
player.color = colors::DARK_RED;
}
|
identifier_body
|
combat.rs
|
pub fn target_tile(tcod: &mut Tcod,
objects: &[Object], game: &mut Game,
max_range: Option<f32>)
-> Option<(i32, i32)> {
use tcod::input::KeyCode::Escape;
loop {
// render the screen. this erases the inventory and shows the names of
// objects under the mouse.
tcod.root.flush();
let event = input::check_for_event(input::KEY_PRESS | input::MOUSE).map(|e| e.1);
let mut key = None;
match event {
Some(Event::Mouse(m)) => tcod.mouse = m,
Some(Event::Key(k)) => key = Some(k),
None =>
|
}
render_all(tcod, objects, game, false);
let (x, y) = (tcod.mouse.cx as i32, tcod.mouse.cy as i32);
// accept the target if the player clicked in FOV, and in case a range
// is specified, if it's in that range
let in_fov = (x < MAP_WIDTH) && (y < MAP_HEIGHT) && tcod.fov.is_in_fov(x, y);
let in_range = max_range.map_or(
true, |range| objects[PLAYER].distance(x, y) <= range);
if tcod.mouse.lbutton_pressed && in_fov && in_range {
return Some((x, y))
}
let escape = key.map_or(false, |k| k.code == Escape);
if tcod.mouse.rbutton_pressed || escape {
return None // cancel if the player right-clicked or pressed Escape
}
}
}
/// find closest enemy, up to a maximum range, and in the player's FOV
pub fn closest_monster(max_range: i32, objects: &mut [Object], tcod: &Tcod) -> Option<usize> {
let mut closest_enemy = None;
let mut closest_dist = (max_range + 1) as f32; // start with (slightly more than) maximum range
for (id, object) in objects.iter().enumerate() {
if (id != PLAYER) && object.fighter.is_some() && object.ai.is_some() &&
tcod.fov.is_in_fov(object.x, object.y)
{
// calculate distance between this object and the player
let dist = objects[PLAYER].distance_to(object);
if dist < closest_dist { // it's closer, so remember it
closest_enemy = Some(id);
closest_dist = dist;
}
}
}
closest_enemy
}
pub fn cast_fireball(_inventory_id: usize, objects: &mut [Object],game: &mut Game, tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target tile to throw a fireball at
game.log.add("Left-click a target tile for the fireball, or right-click to cancel.",
colors::LIGHT_CYAN);
let (x, y) = match target_tile(tcod, objects, game, None) {
Some(tile_pos) => tile_pos,
None => return UseResult::Cancelled,
};
game.log.add(format!("The fireball explodes, burning everything within {} tiles!", FIREBALL_RADIUS),
colors::ORANGE);
let mut xp_to_gain = 0;
for (id, obj) in objects.iter_mut().enumerate() {
if obj.distance(x, y) <= FIREBALL_RADIUS as f32 && obj.fighter.is_some() {
game.log.add(format!("The {} gets burned for {} hit points.", obj.name, FIREBALL_DAMAGE),
colors::ORANGE);
if let Some(xp) = obj.take_damage(FIREBALL_DAMAGE, game) {
// Don't reward the player for burning themself!
if id != PLAYER {
xp_to_gain += xp;
}
}
}
}
objects[PLAYER].fighter.as_mut().unwrap().xp += xp_to_gain;
UseResult::UsedUp
}
pub fn cast_heal(_inventory_id: usize, objects: &mut [Object], game: &mut Game, _tcod: &mut Tcod)
-> UseResult
{
// heal the player
let player = &mut objects[PLAYER];
if let Some(fighter) = player.fighter {
if fighter.hp == player.max_hp(game) {
game.log.add("You are already at full health.", colors::RED);
return UseResult::Cancelled;
}
game.log.add("Your wounds start to feel better!", colors::LIGHT_VIOLET);
player.heal(HEAL_AMOUNT, game);
return UseResult::UsedUp;
}
UseResult::Cancelled
}
pub fn cast_lightning(_inventory_id: usize, objects: &mut [Object], game: &mut Game, tcod: &mut Tcod) -> UseResult
{
// find closest enemy (inside a maximum range) and damage it
let monster_id = closest_monster(LIGHTNING_RANGE, objects, tcod);
if let Some(monster_id) = monster_id {
// zap it!
game.log.add(format!("A lightning bolt strikes the {} with a loud thunder! \
The damage is {} hit points.",
objects[monster_id].name, LIGHTNING_DAMAGE),
colors::LIGHT_BLUE);
objects[monster_id].take_damage(LIGHTNING_DAMAGE, game);
UseResult::UsedUp
} else { // no enemy found within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn cast_confuse(_inventory_id: usize, objects: &mut [Object], game: &mut Game,tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target to confuse
game.log.add("Left-click an enemy to confuse it, or right-click to cancel.", colors::LIGHT_CYAN);
let monster_id = target_monster(tcod, objects, game, Some(CONFUSE_RANGE as f32));
if let Some(monster_id) = monster_id {
let old_ai = objects[monster_id].ai.take().unwrap_or(Ai::Basic);
// replace the monster's AI with a "confused" one; after
// some turns it will restore the old AI
objects[monster_id].ai = Some(Ai::Confused {
previous_ai: Box::new(old_ai),
num_turns: CONFUSE_NUM_TURNS,
});
game.log.add(format!("The eyes of {} look vacant, as he starts to stumble around!",
objects[monster_id].name),
colors::LIGHT_GREEN);
UseResult::UsedUp
} else { // no enemy fonud within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn player_death(player: &mut Object, messages: &mut Messages) {
// the game ended!
// TODO Replace with game.log.add()
message(messages, "You died!", colors::DARK_RED);
// for added effect, transform the player into a corpse!
player.char = CORPSE;
player.color = colors::DARK_RED;
}
pub fn monster_death(monster: &mut Object, messages: &mut Messages) {
// transform it into a nasty corpse! it doesn't block, can't be
// attacked and doesn't move
// TODO Replace with game.log.add()
// message(messages, format!("{} is dead!", monster.name), colors::ORANGE);
message(messages, format!("{} is dead! You gain {} experience points.",
monster.name, monster.fighter.unwrap().xp), colors::ORANGE);
monster.char = CORPSE;
monster.color = colors::DARK_RED;
monster.blocks = false;
monster.fighter = None;
monster.ai = None;
monster.name = format!("remains of {}", monster.name);
}
pub fn player_move_or_attack(dx: i32, dy: i32, objects: &mut [Object], game: &mut Game) {
// the coordinates the player is moving to/attacking
let x = objects[PLAYER].x + dx;
let y = objects[PLAYER].y + dy;
// try to find an attackable object there
let target_id = objects.iter().position(|object| {
object.fighter.is_some() && object.pos() == (x, y)
});
// attack if target found, move otherwise
match target_id {
Some(target_id) => {
let (player, target) = mut_two(PLAYER, target_id, objects);
player.attack(target, game);
}
None => {
move_by(PLAYER, dx, dy, &mut game.map, objects);
}
}
}
pub fn ai_take_turn(monster_id: usize, objects: &mut [Object], game: &mut Game, fov_map: &FovMap) {
use Ai::*;
if let Some(ai) = objects[monster_id].ai.take() {
let new_ai = match ai {
Basic => ai_basic(monster_id, game, objects, fov_map),
Confused{previous_ai, num_turns} => ai_confused(
monster_id, &mut game.map, objects, &mut game.log, previous_ai, num_turns)
};
objects[monster_id].
|
{}
|
conditional_block
|
combat.rs
|
objects: &[Object], game: &mut Game,
max_range: Option<f32>)
-> Option<(i32, i32)> {
use tcod::input::KeyCode::Escape;
loop {
// render the screen. this erases the inventory and shows the names of
// objects under the mouse.
tcod.root.flush();
let event = input::check_for_event(input::KEY_PRESS | input::MOUSE).map(|e| e.1);
let mut key = None;
match event {
Some(Event::Mouse(m)) => tcod.mouse = m,
Some(Event::Key(k)) => key = Some(k),
None => {}
}
render_all(tcod, objects, game, false);
let (x, y) = (tcod.mouse.cx as i32, tcod.mouse.cy as i32);
// accept the target if the player clicked in FOV, and in case a range
// is specified, if it's in that range
let in_fov = (x < MAP_WIDTH) && (y < MAP_HEIGHT) && tcod.fov.is_in_fov(x, y);
let in_range = max_range.map_or(
true, |range| objects[PLAYER].distance(x, y) <= range);
if tcod.mouse.lbutton_pressed && in_fov && in_range {
return Some((x, y))
}
let escape = key.map_or(false, |k| k.code == Escape);
if tcod.mouse.rbutton_pressed || escape {
return None // cancel if the player right-clicked or pressed Escape
}
}
}
/// find closest enemy, up to a maximum range, and in the player's FOV
pub fn closest_monster(max_range: i32, objects: &mut [Object], tcod: &Tcod) -> Option<usize> {
let mut closest_enemy = None;
let mut closest_dist = (max_range + 1) as f32; // start with (slightly more than) maximum range
for (id, object) in objects.iter().enumerate() {
if (id != PLAYER) && object.fighter.is_some() && object.ai.is_some() &&
tcod.fov.is_in_fov(object.x, object.y)
{
// calculate distance between this object and the player
let dist = objects[PLAYER].distance_to(object);
if dist < closest_dist { // it's closer, so remember it
closest_enemy = Some(id);
closest_dist = dist;
}
}
}
closest_enemy
}
pub fn cast_fireball(_inventory_id: usize, objects: &mut [Object],game: &mut Game, tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target tile to throw a fireball at
game.log.add("Left-click a target tile for the fireball, or right-click to cancel.",
colors::LIGHT_CYAN);
let (x, y) = match target_tile(tcod, objects, game, None) {
Some(tile_pos) => tile_pos,
None => return UseResult::Cancelled,
};
game.log.add(format!("The fireball explodes, burning everything within {} tiles!", FIREBALL_RADIUS),
colors::ORANGE);
let mut xp_to_gain = 0;
for (id, obj) in objects.iter_mut().enumerate() {
if obj.distance(x, y) <= FIREBALL_RADIUS as f32 && obj.fighter.is_some() {
game.log.add(format!("The {} gets burned for {} hit points.", obj.name, FIREBALL_DAMAGE),
colors::ORANGE);
if let Some(xp) = obj.take_damage(FIREBALL_DAMAGE, game) {
// Don't reward the player for burning themself!
if id != PLAYER {
xp_to_gain += xp;
}
}
}
}
objects[PLAYER].fighter.as_mut().unwrap().xp += xp_to_gain;
UseResult::UsedUp
}
pub fn cast_heal(_inventory_id: usize, objects: &mut [Object], game: &mut Game, _tcod: &mut Tcod)
-> UseResult
{
// heal the player
let player = &mut objects[PLAYER];
if let Some(fighter) = player.fighter {
if fighter.hp == player.max_hp(game) {
game.log.add("You are already at full health.", colors::RED);
return UseResult::Cancelled;
}
game.log.add("Your wounds start to feel better!", colors::LIGHT_VIOLET);
player.heal(HEAL_AMOUNT, game);
return UseResult::UsedUp;
}
UseResult::Cancelled
}
pub fn cast_lightning(_inventory_id: usize, objects: &mut [Object], game: &mut Game, tcod: &mut Tcod) -> UseResult
{
// find closest enemy (inside a maximum range) and damage it
let monster_id = closest_monster(LIGHTNING_RANGE, objects, tcod);
if let Some(monster_id) = monster_id {
// zap it!
game.log.add(format!("A lightning bolt strikes the {} with a loud thunder! \
The damage is {} hit points.",
objects[monster_id].name, LIGHTNING_DAMAGE),
colors::LIGHT_BLUE);
objects[monster_id].take_damage(LIGHTNING_DAMAGE, game);
UseResult::UsedUp
} else { // no enemy found within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn cast_confuse(_inventory_id: usize, objects: &mut [Object], game: &mut Game,tcod: &mut Tcod)
-> UseResult
{
// ask the player for a target to confuse
game.log.add("Left-click an enemy to confuse it, or right-click to cancel.", colors::LIGHT_CYAN);
let monster_id = target_monster(tcod, objects, game, Some(CONFUSE_RANGE as f32));
if let Some(monster_id) = monster_id {
let old_ai = objects[monster_id].ai.take().unwrap_or(Ai::Basic);
// replace the monster's AI with a "confused" one; after
// some turns it will restore the old AI
objects[monster_id].ai = Some(Ai::Confused {
previous_ai: Box::new(old_ai),
num_turns: CONFUSE_NUM_TURNS,
});
game.log.add(format!("The eyes of {} look vacant, as he starts to stumble around!",
objects[monster_id].name),
colors::LIGHT_GREEN);
UseResult::UsedUp
} else { // no enemy fonud within maximum range
game.log.add("No enemy is close enough to strike.", colors::RED);
UseResult::Cancelled
}
}
pub fn player_death(player: &mut Object, messages: &mut Messages) {
// the game ended!
// TODO Replace with game.log.add()
message(messages, "You died!", colors::DARK_RED);
// for added effect, transform the player into a corpse!
player.char = CORPSE;
player.color = colors::DARK_RED;
}
pub fn monster_death(monster: &mut Object, messages: &mut Messages) {
// transform it into a nasty corpse! it doesn't block, can't be
// attacked and doesn't move
// TODO Replace with game.log.add()
// message(messages, format!("{} is dead!", monster.name), colors::ORANGE);
message(messages, format!("{} is dead! You gain {} experience points.",
monster.name, monster.fighter.unwrap().xp), colors::ORANGE);
monster.char = CORPSE;
monster.color = colors::DARK_RED;
monster.blocks = false;
monster.fighter = None;
monster.ai = None;
monster.name = format!("remains of {}", monster.name);
}
pub fn player_move_or_attack(dx: i32, dy: i32, objects: &mut [Object], game: &mut Game) {
// the coordinates the player is moving to/attacking
let x = objects[PLAYER].x + dx;
let y = objects[PLAYER].y + dy;
// try to find an attackable object there
let target_id = objects.iter().position(|object| {
object.fighter.is_some() && object.pos() == (x, y)
});
// attack if target found, move otherwise
match target_id {
Some(target_id) => {
let (player, target) = mut_two(PLAYER, target_id, objects);
player.attack(target, game);
}
None => {
move_by(PLAYER, dx, dy, &mut game.map, objects);
}
}
}
pub fn ai_take_turn(monster_id: usize, objects: &mut [Object], game: &mut Game, fov_map: &FovMap) {
use Ai::*;
if let Some(ai) = objects[monster_id].ai.take() {
let new_ai = match ai {
Basic => ai_basic(monster_id, game, objects, fov_map),
Confused{previous_ai, num_turns} => ai_confused(
monster_id, &mut game.map, objects, &mut game.log, previous_ai, num_turns)
};
objects[monster_id].ai = Some(new_ai);
}
}
pub fn
|
ai_basic
|
identifier_name
|
|
ikdbtest_gl.py
|
(GLPluginProgram):
def __init__(self,world,name):
GLPluginProgram.__init__(self,name)
self.widgetPlugin = GLWidgetPlugin()
self.setPlugin(self.widgetPlugin)
self.widgetMaster = self.widgetPlugin.klamptwidgetmaster
self.world = world
def display(self):
GLPluginProgram.display(self)
self.world.drawGL()
else:
NEW_KLAMPT = False
from klampt import ik,loader
from klampt.glrobotprogram import *
from klampt import PointPoser,TransformPoser
from klampt import robotcollide as collide
from klampt import se3
import sys
import traceback
import numpy as np
#preload
from sklearn.neighbors import NearestNeighbors,BallTree
class IKDBVisualTester(GLWidgetProgram):
def __init__(self,visWorld,planningWorld,name="IK Database visual tester"):
GLWidgetProgram.__init__(self,visWorld,name)
self.planningWorld = planningWorld
self.collider = collide.WorldCollider(visWorld)
self.ikdb = ManagedIKDatabase(planningWorld.robot(0))
self.ikWidgets = []
self.ikIndices = []
self.ikProblem = IKProblem()
self.ikProblem.setFeasibilityTest('collisionFree',None)
qmin,qmax = planningWorld.robot(0).getJointLimits()
self.ikProblem.setCostFunction('jointRangeCost_dynamic',[qmin,qmax])
self.drawDb = False
self.continuous = False
self.reSolve = False
self.currentConfig = self.world.robot(0).getConfig()
def mousefunc(self,button,state,x,y):
#Put your mouse handler here
#the current example prints out the list of objects clicked whenever
#you right click
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.reSolve = False
dragging = False
if NEW_KLAMPT:
dragging = self.widgetPlugin.klamptwidgetdragging
else:
dragging = self.draggingWidget
if not dragging and button == 2 and state==0:
#down
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
lpt = se3.apply(se3.inv(link.getTransform()),wpt)
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(PointPoser())
self.ikWidgets[-1].set(wpt)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,local=lpt,world=wpt))
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.refresh()
return
#the dx,dy arguments are needed to be cross-compatible between 0.6.x and 0.7
def motionfunc(self,x,y,dx=0,dy=0):
dragging = False
if NEW_KLAMPT:
retval = GLWidgetProgram.motionfunc(self,x,y,dx,dy)
dragging = self.widgetPlugin.klamptwidgetdragging
else:
retval = GLWidgetProgram.motionfunc(self,x,y)
dragging = self.draggingWidget
if dragging:
#update all the IK objectives
for i in range(len(self.ikWidgets)):
index = self.ikIndices[i]
if isinstance(self.ikWidgets[i],PointPoser):
wptnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
lpt,wptold = obj.getPosition()
obj.setFixedPoint(link,lpt,wptnew)
#don't solve now, wait for refresh to process
if self.continuous and wptnew != wptold:
self.reSolve = True
elif isinstance(self.ikWidgets[i],TransformPoser):
Rnew,tnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
Rold,told = obj.getTransform()
obj.setFixedTransform(link,Rnew,tnew)
#don't solve now, wait for refresh to process
if self.continuous and (Rnew,tnew) != (Rold,told):
self.reSolve = True
return retval
def keyboardfunc(self,c,x,y):
if c=='h':
|
print ('[space]: tests the current configuration')
print ('d: deletes IK constraint')
print ('t: adds a new rotation-fixed IK constraint')
print ('f: flushes the current database to disk')
print ('s: saves the current database to disk')
print ('b: performs one background step')
print ('B: starts / stops the background thread')
print ('v: toggles display of the database')
print ('c: toggles continuous re-solving of IK constraint its as being moved')
print ('o: toggles soft / hard IK constraints')
elif c==' ':
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
print ("Solved")
self.currentConfig = soln
self.refresh()
else:
print ("Failure")
elif c=='d':
for i,w in enumerate(self.ikWidgets):
if w.hasHighlight():
print ("Deleting IK widget")
#delete it
index = self.ikIndices[i]
self.widgetMaster.remove(w)
del self.ikWidgets[i]
del self.ikIndices[i]
del self.ikProblem.objectives[index]
for j in range(len(self.ikIndices)):
self.ikIndices[j] = j
self.refresh()
break
elif c=='t':
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
Tlink = link.getTransform()
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(TransformPoser())
self.ikWidgets[-1].set(*Tlink)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,R=Tlink[0],t=Tlink[1]))
self.refresh()
elif c=='f':
self.ikdb.flush()
elif c=='s':
self.ikdb.save()
elif c=='b':
self.ikdb.backgroundStep()
self.refresh()
elif c=='B':
if hasattr(self.ikdb,'thread'):
self.ikdb.stopBackgroundLoop()
else:
self.ikdb.startBackgroundLoop(0)
elif c=='v':
self.drawDb = not self.drawDb
elif c=='c':
self.continuous = not self.continuous
elif c == 'o':
self.ikProblem.setSoftObjectives(not self.ikProblem.softObjectives)
def display(self):
if self.reSolve:
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
self.currentConfig = soln
self.reSolve = False
self.world.robot(0).setConfig(self.currentConfig)
GLWidgetProgram.display(self)
glDisable(GL_LIGHTING)
#draw IK goals
for obj in self.ikProblem.objectives:
linkindex = obj.link()
link = self.world.robot(0).link(linkindex)
lp,wpdes = obj.getPosition()
wp = se3.apply(link.getTransform(),lp)
glLineWidth(4.0)
glDisable(GL_LIGHTING)
glColor3f(0,1,0)
glBegin(GL_LINES)
glVertex3f(*wp)
glVertex3f(*wpdes)
glEnd()
glLineWidth(1)
#draw end positions of solved problems
if self.drawDb:
glPointSize(3.0)
glBegin(GL_POINTS)
for k,db in iteritems(self.ikdb.databases):
for i in range(db.numProblems()):
try:
p = db.getProblem(i)
except Exception as e:
traceback.print_exc()
exit(0)
if db.solutions[i] is None:
glColor3f(1,0,0)
else:
glColor3f(0,0,1)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glColor3f(1,1,0)
for pjson,soln in self.ikdb.backburner:
p = IKProblem()
p.fromJson(pjson)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glEnd()
return
def click_world(self,x,y):
"""Helper: returns (obj,pt) where obj is the closest world object
clicked, and pt is the associated clicked point (in world coordinates).
If no point is clicked, returns None."""
#get the viewport ray
if NEW_KLAMPT:
(s,d) =
|
print ('HELP:')
print ('[right-click]: add a new IK constraint')
|
random_line_split
|
ikdbtest_gl.py
|
__init__(self,world,name):
GLPluginProgram.__init__(self,name)
self.widgetPlugin = GLWidgetPlugin()
self.setPlugin(self.widgetPlugin)
self.widgetMaster = self.widgetPlugin.klamptwidgetmaster
self.world = world
def display(self):
GLPluginProgram.display(self)
self.world.drawGL()
else:
NEW_KLAMPT = False
from klampt import ik,loader
from klampt.glrobotprogram import *
from klampt import PointPoser,TransformPoser
from klampt import robotcollide as collide
from klampt import se3
import sys
import traceback
import numpy as np
#preload
from sklearn.neighbors import NearestNeighbors,BallTree
class IKDBVisualTester(GLWidgetProgram):
def __init__(self,visWorld,planningWorld,name="IK Database visual tester"):
GLWidgetProgram.__init__(self,visWorld,name)
self.planningWorld = planningWorld
self.collider = collide.WorldCollider(visWorld)
self.ikdb = ManagedIKDatabase(planningWorld.robot(0))
self.ikWidgets = []
self.ikIndices = []
self.ikProblem = IKProblem()
self.ikProblem.setFeasibilityTest('collisionFree',None)
qmin,qmax = planningWorld.robot(0).getJointLimits()
self.ikProblem.setCostFunction('jointRangeCost_dynamic',[qmin,qmax])
self.drawDb = False
self.continuous = False
self.reSolve = False
self.currentConfig = self.world.robot(0).getConfig()
def mousefunc(self,button,state,x,y):
#Put your mouse handler here
#the current example prints out the list of objects clicked whenever
#you right click
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.reSolve = False
dragging = False
if NEW_KLAMPT:
dragging = self.widgetPlugin.klamptwidgetdragging
else:
dragging = self.draggingWidget
if not dragging and button == 2 and state==0:
#down
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
lpt = se3.apply(se3.inv(link.getTransform()),wpt)
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(PointPoser())
self.ikWidgets[-1].set(wpt)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,local=lpt,world=wpt))
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.refresh()
return
#the dx,dy arguments are needed to be cross-compatible between 0.6.x and 0.7
def motionfunc(self,x,y,dx=0,dy=0):
dragging = False
if NEW_KLAMPT:
retval = GLWidgetProgram.motionfunc(self,x,y,dx,dy)
dragging = self.widgetPlugin.klamptwidgetdragging
else:
retval = GLWidgetProgram.motionfunc(self,x,y)
dragging = self.draggingWidget
if dragging:
#update all the IK objectives
for i in range(len(self.ikWidgets)):
index = self.ikIndices[i]
if isinstance(self.ikWidgets[i],PointPoser):
wptnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
lpt,wptold = obj.getPosition()
obj.setFixedPoint(link,lpt,wptnew)
#don't solve now, wait for refresh to process
if self.continuous and wptnew != wptold:
self.reSolve = True
elif isinstance(self.ikWidgets[i],TransformPoser):
Rnew,tnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
Rold,told = obj.getTransform()
obj.setFixedTransform(link,Rnew,tnew)
#don't solve now, wait for refresh to process
if self.continuous and (Rnew,tnew) != (Rold,told):
self.reSolve = True
return retval
def keyboardfunc(self,c,x,y):
if c=='h':
print ('HELP:')
print ('[right-click]: add a new IK constraint')
print ('[space]: tests the current configuration')
print ('d: deletes IK constraint')
print ('t: adds a new rotation-fixed IK constraint')
print ('f: flushes the current database to disk')
print ('s: saves the current database to disk')
print ('b: performs one background step')
print ('B: starts / stops the background thread')
print ('v: toggles display of the database')
print ('c: toggles continuous re-solving of IK constraint its as being moved')
print ('o: toggles soft / hard IK constraints')
elif c==' ':
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
print ("Solved")
self.currentConfig = soln
self.refresh()
else:
print ("Failure")
elif c=='d':
for i,w in enumerate(self.ikWidgets):
if w.hasHighlight():
print ("Deleting IK widget")
#delete it
index = self.ikIndices[i]
self.widgetMaster.remove(w)
del self.ikWidgets[i]
del self.ikIndices[i]
del self.ikProblem.objectives[index]
for j in range(len(self.ikIndices)):
self.ikIndices[j] = j
self.refresh()
break
elif c=='t':
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
Tlink = link.getTransform()
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(TransformPoser())
self.ikWidgets[-1].set(*Tlink)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,R=Tlink[0],t=Tlink[1]))
self.refresh()
elif c=='f':
self.ikdb.flush()
elif c=='s':
self.ikdb.save()
elif c=='b':
self.ikdb.backgroundStep()
self.refresh()
elif c=='B':
if hasattr(self.ikdb,'thread'):
self.ikdb.stopBackgroundLoop()
else:
self.ikdb.startBackgroundLoop(0)
elif c=='v':
self.drawDb = not self.drawDb
elif c=='c':
self.continuous = not self.continuous
elif c == 'o':
self.ikProblem.setSoftObjectives(not self.ikProblem.softObjectives)
def display(self):
if self.reSolve:
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
self.currentConfig = soln
self.reSolve = False
self.world.robot(0).setConfig(self.currentConfig)
GLWidgetProgram.display(self)
glDisable(GL_LIGHTING)
#draw IK goals
for obj in self.ikProblem.objectives:
linkindex = obj.link()
link = self.world.robot(0).link(linkindex)
lp,wpdes = obj.getPosition()
wp = se3.apply(link.getTransform(),lp)
glLineWidth(4.0)
glDisable(GL_LIGHTING)
glColor3f(0,1,0)
glBegin(GL_LINES)
glVertex3f(*wp)
glVertex3f(*wpdes)
glEnd()
glLineWidth(1)
#draw end positions of solved problems
if self.drawDb:
glPointSize(3.0)
glBegin(GL_POINTS)
for k,db in iteritems(self.ikdb.databases):
for i in range(db.numProblems()):
try:
p = db.getProblem(i)
except Exception as e:
traceback.print_exc()
exit(0)
if db.solutions[i] is None:
glColor3f(1,0,0)
else:
glColor3f(0,0,1)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glColor3f(1,1,0)
for pjson,soln in self.ikdb.backburner:
p = IKProblem()
p.fromJson(pjson)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glEnd()
return
def click_world(self,x,y):
"""Helper: returns (obj,pt) where obj is the closest world object
clicked, and pt is the associated clicked point (in world coordinates).
If no point is clicked, returns None."""
#get the viewport ray
if NEW_KLAMPT:
|
(s,d) = self.view.click_ray(x,y)
|
conditional_block
|
|
ikdbtest_gl.py
|
PluginProgram):
def __init__(self,world,name):
GLPluginProgram.__init__(self,name)
self.widgetPlugin = GLWidgetPlugin()
self.setPlugin(self.widgetPlugin)
self.widgetMaster = self.widgetPlugin.klamptwidgetmaster
self.world = world
def display(self):
GLPluginProgram.display(self)
self.world.drawGL()
else:
NEW_KLAMPT = False
from klampt import ik,loader
from klampt.glrobotprogram import *
from klampt import PointPoser,TransformPoser
from klampt import robotcollide as collide
from klampt import se3
import sys
import traceback
import numpy as np
#preload
from sklearn.neighbors import NearestNeighbors,BallTree
class IKDBVisualTester(GLWidgetProgram):
def
|
(self,visWorld,planningWorld,name="IK Database visual tester"):
GLWidgetProgram.__init__(self,visWorld,name)
self.planningWorld = planningWorld
self.collider = collide.WorldCollider(visWorld)
self.ikdb = ManagedIKDatabase(planningWorld.robot(0))
self.ikWidgets = []
self.ikIndices = []
self.ikProblem = IKProblem()
self.ikProblem.setFeasibilityTest('collisionFree',None)
qmin,qmax = planningWorld.robot(0).getJointLimits()
self.ikProblem.setCostFunction('jointRangeCost_dynamic',[qmin,qmax])
self.drawDb = False
self.continuous = False
self.reSolve = False
self.currentConfig = self.world.robot(0).getConfig()
def mousefunc(self,button,state,x,y):
#Put your mouse handler here
#the current example prints out the list of objects clicked whenever
#you right click
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.reSolve = False
dragging = False
if NEW_KLAMPT:
dragging = self.widgetPlugin.klamptwidgetdragging
else:
dragging = self.draggingWidget
if not dragging and button == 2 and state==0:
#down
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
lpt = se3.apply(se3.inv(link.getTransform()),wpt)
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(PointPoser())
self.ikWidgets[-1].set(wpt)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,local=lpt,world=wpt))
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.refresh()
return
#the dx,dy arguments are needed to be cross-compatible between 0.6.x and 0.7
def motionfunc(self,x,y,dx=0,dy=0):
dragging = False
if NEW_KLAMPT:
retval = GLWidgetProgram.motionfunc(self,x,y,dx,dy)
dragging = self.widgetPlugin.klamptwidgetdragging
else:
retval = GLWidgetProgram.motionfunc(self,x,y)
dragging = self.draggingWidget
if dragging:
#update all the IK objectives
for i in range(len(self.ikWidgets)):
index = self.ikIndices[i]
if isinstance(self.ikWidgets[i],PointPoser):
wptnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
lpt,wptold = obj.getPosition()
obj.setFixedPoint(link,lpt,wptnew)
#don't solve now, wait for refresh to process
if self.continuous and wptnew != wptold:
self.reSolve = True
elif isinstance(self.ikWidgets[i],TransformPoser):
Rnew,tnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
Rold,told = obj.getTransform()
obj.setFixedTransform(link,Rnew,tnew)
#don't solve now, wait for refresh to process
if self.continuous and (Rnew,tnew) != (Rold,told):
self.reSolve = True
return retval
def keyboardfunc(self,c,x,y):
if c=='h':
print ('HELP:')
print ('[right-click]: add a new IK constraint')
print ('[space]: tests the current configuration')
print ('d: deletes IK constraint')
print ('t: adds a new rotation-fixed IK constraint')
print ('f: flushes the current database to disk')
print ('s: saves the current database to disk')
print ('b: performs one background step')
print ('B: starts / stops the background thread')
print ('v: toggles display of the database')
print ('c: toggles continuous re-solving of IK constraint its as being moved')
print ('o: toggles soft / hard IK constraints')
elif c==' ':
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
print ("Solved")
self.currentConfig = soln
self.refresh()
else:
print ("Failure")
elif c=='d':
for i,w in enumerate(self.ikWidgets):
if w.hasHighlight():
print ("Deleting IK widget")
#delete it
index = self.ikIndices[i]
self.widgetMaster.remove(w)
del self.ikWidgets[i]
del self.ikIndices[i]
del self.ikProblem.objectives[index]
for j in range(len(self.ikIndices)):
self.ikIndices[j] = j
self.refresh()
break
elif c=='t':
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
Tlink = link.getTransform()
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(TransformPoser())
self.ikWidgets[-1].set(*Tlink)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,R=Tlink[0],t=Tlink[1]))
self.refresh()
elif c=='f':
self.ikdb.flush()
elif c=='s':
self.ikdb.save()
elif c=='b':
self.ikdb.backgroundStep()
self.refresh()
elif c=='B':
if hasattr(self.ikdb,'thread'):
self.ikdb.stopBackgroundLoop()
else:
self.ikdb.startBackgroundLoop(0)
elif c=='v':
self.drawDb = not self.drawDb
elif c=='c':
self.continuous = not self.continuous
elif c == 'o':
self.ikProblem.setSoftObjectives(not self.ikProblem.softObjectives)
def display(self):
if self.reSolve:
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
self.currentConfig = soln
self.reSolve = False
self.world.robot(0).setConfig(self.currentConfig)
GLWidgetProgram.display(self)
glDisable(GL_LIGHTING)
#draw IK goals
for obj in self.ikProblem.objectives:
linkindex = obj.link()
link = self.world.robot(0).link(linkindex)
lp,wpdes = obj.getPosition()
wp = se3.apply(link.getTransform(),lp)
glLineWidth(4.0)
glDisable(GL_LIGHTING)
glColor3f(0,1,0)
glBegin(GL_LINES)
glVertex3f(*wp)
glVertex3f(*wpdes)
glEnd()
glLineWidth(1)
#draw end positions of solved problems
if self.drawDb:
glPointSize(3.0)
glBegin(GL_POINTS)
for k,db in iteritems(self.ikdb.databases):
for i in range(db.numProblems()):
try:
p = db.getProblem(i)
except Exception as e:
traceback.print_exc()
exit(0)
if db.solutions[i] is None:
glColor3f(1,0,0)
else:
glColor3f(0,0,1)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glColor3f(1,1,0)
for pjson,soln in self.ikdb.backburner:
p = IKProblem()
p.fromJson(pjson)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glEnd()
return
def click_world(self,x,y):
"""Helper: returns (obj,pt) where obj is the closest world object
clicked, and pt is the associated clicked point (in world coordinates).
If no point is clicked, returns None."""
#get the viewport ray
if NEW_KLAMPT:
(s,d) =
|
__init__
|
identifier_name
|
ikdbtest_gl.py
|
PluginProgram):
def __init__(self,world,name):
GLPluginProgram.__init__(self,name)
self.widgetPlugin = GLWidgetPlugin()
self.setPlugin(self.widgetPlugin)
self.widgetMaster = self.widgetPlugin.klamptwidgetmaster
self.world = world
def display(self):
GLPluginProgram.display(self)
self.world.drawGL()
else:
NEW_KLAMPT = False
from klampt import ik,loader
from klampt.glrobotprogram import *
from klampt import PointPoser,TransformPoser
from klampt import robotcollide as collide
from klampt import se3
import sys
import traceback
import numpy as np
#preload
from sklearn.neighbors import NearestNeighbors,BallTree
class IKDBVisualTester(GLWidgetProgram):
def __init__(self,visWorld,planningWorld,name="IK Database visual tester"):
|
def mousefunc(self,button,state,x,y):
#Put your mouse handler here
#the current example prints out the list of objects clicked whenever
#you right click
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.reSolve = False
dragging = False
if NEW_KLAMPT:
dragging = self.widgetPlugin.klamptwidgetdragging
else:
dragging = self.draggingWidget
if not dragging and button == 2 and state==0:
#down
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
lpt = se3.apply(se3.inv(link.getTransform()),wpt)
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(PointPoser())
self.ikWidgets[-1].set(wpt)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,local=lpt,world=wpt))
GLWidgetProgram.mousefunc(self,button,state,x,y)
self.refresh()
return
#the dx,dy arguments are needed to be cross-compatible between 0.6.x and 0.7
def motionfunc(self,x,y,dx=0,dy=0):
dragging = False
if NEW_KLAMPT:
retval = GLWidgetProgram.motionfunc(self,x,y,dx,dy)
dragging = self.widgetPlugin.klamptwidgetdragging
else:
retval = GLWidgetProgram.motionfunc(self,x,y)
dragging = self.draggingWidget
if dragging:
#update all the IK objectives
for i in range(len(self.ikWidgets)):
index = self.ikIndices[i]
if isinstance(self.ikWidgets[i],PointPoser):
wptnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
lpt,wptold = obj.getPosition()
obj.setFixedPoint(link,lpt,wptnew)
#don't solve now, wait for refresh to process
if self.continuous and wptnew != wptold:
self.reSolve = True
elif isinstance(self.ikWidgets[i],TransformPoser):
Rnew,tnew = self.ikWidgets[i].get()
obj = self.ikProblem.objectives[index]
link = obj.link()
Rold,told = obj.getTransform()
obj.setFixedTransform(link,Rnew,tnew)
#don't solve now, wait for refresh to process
if self.continuous and (Rnew,tnew) != (Rold,told):
self.reSolve = True
return retval
def keyboardfunc(self,c,x,y):
if c=='h':
print ('HELP:')
print ('[right-click]: add a new IK constraint')
print ('[space]: tests the current configuration')
print ('d: deletes IK constraint')
print ('t: adds a new rotation-fixed IK constraint')
print ('f: flushes the current database to disk')
print ('s: saves the current database to disk')
print ('b: performs one background step')
print ('B: starts / stops the background thread')
print ('v: toggles display of the database')
print ('c: toggles continuous re-solving of IK constraint its as being moved')
print ('o: toggles soft / hard IK constraints')
elif c==' ':
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
print ("Solved")
self.currentConfig = soln
self.refresh()
else:
print ("Failure")
elif c=='d':
for i,w in enumerate(self.ikWidgets):
if w.hasHighlight():
print ("Deleting IK widget")
#delete it
index = self.ikIndices[i]
self.widgetMaster.remove(w)
del self.ikWidgets[i]
del self.ikIndices[i]
del self.ikProblem.objectives[index]
for j in range(len(self.ikIndices)):
self.ikIndices[j] = j
self.refresh()
break
elif c=='t':
clicked = self.click_world(x,y)
if clicked is not None and isinstance(clicked[0],RobotModelLink):
#make a new widget
link, wpt = clicked
Tlink = link.getTransform()
self.ikIndices.append(len(self.ikWidgets))
self.ikWidgets.append(TransformPoser())
self.ikWidgets[-1].set(*Tlink)
self.widgetMaster.add(self.ikWidgets[-1])
self.ikProblem.addObjective(ik.objective(link,R=Tlink[0],t=Tlink[1]))
self.refresh()
elif c=='f':
self.ikdb.flush()
elif c=='s':
self.ikdb.save()
elif c=='b':
self.ikdb.backgroundStep()
self.refresh()
elif c=='B':
if hasattr(self.ikdb,'thread'):
self.ikdb.stopBackgroundLoop()
else:
self.ikdb.startBackgroundLoop(0)
elif c=='v':
self.drawDb = not self.drawDb
elif c=='c':
self.continuous = not self.continuous
elif c == 'o':
self.ikProblem.setSoftObjectives(not self.ikProblem.softObjectives)
def display(self):
if self.reSolve:
self.planningWorld.robot(0).setConfig(self.currentConfig)
soln = self.ikdb.solve(self.ikProblem)
if soln:
self.currentConfig = soln
self.reSolve = False
self.world.robot(0).setConfig(self.currentConfig)
GLWidgetProgram.display(self)
glDisable(GL_LIGHTING)
#draw IK goals
for obj in self.ikProblem.objectives:
linkindex = obj.link()
link = self.world.robot(0).link(linkindex)
lp,wpdes = obj.getPosition()
wp = se3.apply(link.getTransform(),lp)
glLineWidth(4.0)
glDisable(GL_LIGHTING)
glColor3f(0,1,0)
glBegin(GL_LINES)
glVertex3f(*wp)
glVertex3f(*wpdes)
glEnd()
glLineWidth(1)
#draw end positions of solved problems
if self.drawDb:
glPointSize(3.0)
glBegin(GL_POINTS)
for k,db in iteritems(self.ikdb.databases):
for i in range(db.numProblems()):
try:
p = db.getProblem(i)
except Exception as e:
traceback.print_exc()
exit(0)
if db.solutions[i] is None:
glColor3f(1,0,0)
else:
glColor3f(0,0,1)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glColor3f(1,1,0)
for pjson,soln in self.ikdb.backburner:
p = IKProblem()
p.fromJson(pjson)
for obj in p.objectives:
lp,wpdes = obj.getPosition()
glVertex3f(*wpdes)
glEnd()
return
def click_world(self,x,y):
"""Helper: returns (obj,pt) where obj is the closest world object
clicked, and pt is the associated clicked point (in world coordinates).
If no point is clicked, returns None."""
#get the viewport ray
if NEW_KLAMPT:
(s,d) =
|
GLWidgetProgram.__init__(self,visWorld,name)
self.planningWorld = planningWorld
self.collider = collide.WorldCollider(visWorld)
self.ikdb = ManagedIKDatabase(planningWorld.robot(0))
self.ikWidgets = []
self.ikIndices = []
self.ikProblem = IKProblem()
self.ikProblem.setFeasibilityTest('collisionFree',None)
qmin,qmax = planningWorld.robot(0).getJointLimits()
self.ikProblem.setCostFunction('jointRangeCost_dynamic',[qmin,qmax])
self.drawDb = False
self.continuous = False
self.reSolve = False
self.currentConfig = self.world.robot(0).getConfig()
|
identifier_body
|
index.go
|
can
// only happen for line numbers; give it no line number (= 0)
x = 0
}
// encode kind: bits [1..4)
x |= SpotInfo(kind) << 1;
// encode isIndex: bit 0
if isIndex {
x |= 1
}
return x;
}
func (x SpotInfo) Kind() SpotKind { return SpotKind(x >> 1 & 7) }
func (x SpotInfo) Lori() int { return int(x >> 4) }
func (x SpotInfo) IsIndex() bool { return x&1 != 0 }
// ----------------------------------------------------------------------------
// KindRun
// Debugging support. Disable to see multiple entries per line.
const removeDuplicates = true
// A KindRun is a run of SpotInfos of the same kind in a given file.
type KindRun struct {
Kind SpotKind;
Infos []SpotInfo;
}
// KindRuns are sorted by line number or index. Since the isIndex bit
// is always the same for all infos in one list we can compare lori's.
func (f *KindRun) Len() int { return len(f.Infos) }
func (f *KindRun) Less(i, j int) bool { return f.Infos[i].Lori() < f.Infos[j].Lori() }
func (f *KindRun) Swap(i, j int) { f.Infos[i], f.Infos[j] = f.Infos[j], f.Infos[i] }
// FileRun contents are sorted by Kind for the reduction into KindRuns.
func lessKind(x, y interface{}) bool { return x.(SpotInfo).Kind() < y.(SpotInfo).Kind() }
// newKindRun allocates a new KindRun from the SpotInfo run [i, j) in h.
func newKindRun(h *RunList, i, j int) interface{} {
kind := h.At(i).(SpotInfo).Kind();
infos := make([]SpotInfo, j-i);
k := 0;
for ; i < j; i++ {
infos[k] = h.At(i).(SpotInfo);
k++;
}
run := &KindRun{kind, infos};
// Spots were sorted by file and kind to create this run.
// Within this run, sort them by line number or index.
sort.Sort(run);
if removeDuplicates {
// Since both the lori and kind field must be
// same for duplicates, and since the isIndex
// bit is always the same for all infos in one
// list we can simply compare the entire info.
k := 0;
var prev SpotInfo;
for i, x := range infos {
if x != prev || i == 0 {
infos[k] = x;
k++;
prev = x;
}
}
run.Infos = infos[0:k];
}
return run;
}
// ----------------------------------------------------------------------------
// FileRun
// A Pak describes a Go package.
type Pak struct {
Path string; // path of directory containing the package
Name string; // package name as declared by package clause
}
// Paks are sorted by name (primary key) and by import path (secondary key).
func (p *Pak) less(q *Pak) bool {
return p.Name < q.Name || p.Name == q.Name && p.Path < q.Path
}
// A File describes a Go file.
type File struct {
Path string; // complete file name
Pak Pak; // the package to which the file belongs
}
// A Spot describes a single occurence of a word.
type Spot struct {
File *File;
Info SpotInfo;
}
// A FileRun is a list of KindRuns belonging to the same file.
type FileRun struct {
File *File;
Groups []*KindRun;
}
// Spots are sorted by path for the reduction into FileRuns.
func lessSpot(x, y interface{}) bool { return x.(Spot).File.Path < y.(Spot).File.Path }
// newFileRun allocates a new FileRun from the Spot run [i, j) in h.
func newFileRun(h0 *RunList, i, j int) interface{} {
file := h0.At(i).(Spot).File;
// reduce the list of Spots into a list of KindRuns
var h1 RunList;
h1.Vector.Init(j - i);
k := 0;
for ; i < j; i++ {
h1.Set(k, h0.At(i).(Spot).Info);
k++;
}
h2 := h1.reduce(lessKind, newKindRun);
// create the FileRun
groups := make([]*KindRun, h2.Len());
for i := 0; i < h2.Len(); i++ {
groups[i] = h2.At(i).(*KindRun)
}
return &FileRun{file, groups};
}
// ----------------------------------------------------------------------------
// PakRun
// A PakRun describes a run of *FileRuns of a package.
type PakRun struct {
Pak Pak;
Files []*FileRun;
}
// Sorting support for files within a PakRun.
func (p *PakRun) Len() int { return len(p.Files) }
func (p *PakRun) Less(i, j int) bool { return p.Files[i].File.Path < p.Files[j].File.Path }
func (p *PakRun) Swap(i, j int) { p.Files[i], p.Files[j] = p.Files[j], p.Files[i] }
// FileRuns are sorted by package for the reduction into PakRuns.
func lessFileRun(x, y interface{}) bool {
return x.(*FileRun).File.Pak.less(&y.(*FileRun).File.Pak)
}
// newPakRun allocates a new PakRun from the *FileRun run [i, j) in h.
func newPakRun(h *RunList, i, j int) interface{} {
pak := h.At(i).(*FileRun).File.Pak;
files := make([]*FileRun, j-i);
k := 0;
for ; i < j; i++
|
run := &PakRun{pak, files};
sort.Sort(run); // files were sorted by package; sort them by file now
return run;
}
// ----------------------------------------------------------------------------
// HitList
// A HitList describes a list of PakRuns.
type HitList []*PakRun
// PakRuns are sorted by package.
func lessPakRun(x, y interface{}) bool { return x.(*PakRun).Pak.less(&y.(*PakRun).Pak) }
func reduce(h0 *RunList) HitList {
// reduce a list of Spots into a list of FileRuns
h1 := h0.reduce(lessSpot, newFileRun);
// reduce a list of FileRuns into a list of PakRuns
h2 := h1.reduce(lessFileRun, newPakRun);
// sort the list of PakRuns by package
h2.sort(lessPakRun);
// create a HitList
h := make(HitList, h2.Len());
for i := 0; i < h2.Len(); i++ {
h[i] = h2.At(i).(*PakRun)
}
return h;
}
func (h HitList) filter(pakname string) HitList {
// determine number of matching packages (most of the time just one)
n := 0;
for _, p := range h {
if p.Pak.Name == pakname {
n++
}
}
// create filtered HitList
hh := make(HitList, n);
i := 0;
for _, p := range h {
if p.Pak.Name == pakname {
hh[i] = p;
i++;
}
}
return hh;
}
// ----------------------------------------------------------------------------
// AltWords
type wordPair struct {
canon string; // canonical word spelling (all lowercase)
alt string; // alternative spelling
}
// An AltWords describes a list of alternative spellings for a
// canonical (all lowercase) spelling of a word.
type AltWords struct {
Canon string; // canonical word spelling (all lowercase)
Alts []string; // alternative spelling for the same word
}
// wordPairs are sorted by their canonical spelling.
func lessWordPair(x, y interface{}) bool { return x.(*wordPair).canon < y.(*wordPair).canon }
// newAltWords allocates a new AltWords from the *wordPair run [i, j) in h.
func newAltWords(h *RunList, i, j int) interface{} {
canon := h.At(i).(*wordPair).canon;
alts := make([]string, j-i);
k := 0;
for ; i < j; i++ {
alts[k] = h.At(i).(*wordPair).alt;
k++;
}
return &AltWords{canon, alts};
}
func (a *AltWords) filter(s string) *AltWords {
if len(a.Alts) == 1 && a.Alts[0] == s {
// there are no different alternatives
return nil
}
// make a new AltWords with the current spelling removed
alts := make([]string, len(a.Alts));
i := 0;
for _, w := range a.Alts {
|
{
files[k] = h.At(i).(*FileRun);
k++;
}
|
conditional_block
|
index.go
|
can
// only happen for line numbers; give it no line number (= 0)
x = 0
}
// encode kind: bits [1..4)
x |= SpotInfo(kind) << 1;
// encode isIndex: bit 0
if isIndex {
x |= 1
}
return x;
}
func (x SpotInfo) Kind() SpotKind
|
func (x SpotInfo) Lori() int { return int(x >> 4) }
func (x SpotInfo) IsIndex() bool { return x&1 != 0 }
// ----------------------------------------------------------------------------
// KindRun
// Debugging support. Disable to see multiple entries per line.
const removeDuplicates = true
// A KindRun is a run of SpotInfos of the same kind in a given file.
type KindRun struct {
Kind SpotKind;
Infos []SpotInfo;
}
// KindRuns are sorted by line number or index. Since the isIndex bit
// is always the same for all infos in one list we can compare lori's.
func (f *KindRun) Len() int { return len(f.Infos) }
func (f *KindRun) Less(i, j int) bool { return f.Infos[i].Lori() < f.Infos[j].Lori() }
func (f *KindRun) Swap(i, j int) { f.Infos[i], f.Infos[j] = f.Infos[j], f.Infos[i] }
// FileRun contents are sorted by Kind for the reduction into KindRuns.
func lessKind(x, y interface{}) bool { return x.(SpotInfo).Kind() < y.(SpotInfo).Kind() }
// newKindRun allocates a new KindRun from the SpotInfo run [i, j) in h.
func newKindRun(h *RunList, i, j int) interface{} {
kind := h.At(i).(SpotInfo).Kind();
infos := make([]SpotInfo, j-i);
k := 0;
for ; i < j; i++ {
infos[k] = h.At(i).(SpotInfo);
k++;
}
run := &KindRun{kind, infos};
// Spots were sorted by file and kind to create this run.
// Within this run, sort them by line number or index.
sort.Sort(run);
if removeDuplicates {
// Since both the lori and kind field must be
// same for duplicates, and since the isIndex
// bit is always the same for all infos in one
// list we can simply compare the entire info.
k := 0;
var prev SpotInfo;
for i, x := range infos {
if x != prev || i == 0 {
infos[k] = x;
k++;
prev = x;
}
}
run.Infos = infos[0:k];
}
return run;
}
// ----------------------------------------------------------------------------
// FileRun
// A Pak describes a Go package.
type Pak struct {
Path string; // path of directory containing the package
Name string; // package name as declared by package clause
}
// Paks are sorted by name (primary key) and by import path (secondary key).
func (p *Pak) less(q *Pak) bool {
return p.Name < q.Name || p.Name == q.Name && p.Path < q.Path
}
// A File describes a Go file.
type File struct {
Path string; // complete file name
Pak Pak; // the package to which the file belongs
}
// A Spot describes a single occurence of a word.
type Spot struct {
File *File;
Info SpotInfo;
}
// A FileRun is a list of KindRuns belonging to the same file.
type FileRun struct {
File *File;
Groups []*KindRun;
}
// Spots are sorted by path for the reduction into FileRuns.
func lessSpot(x, y interface{}) bool { return x.(Spot).File.Path < y.(Spot).File.Path }
// newFileRun allocates a new FileRun from the Spot run [i, j) in h.
func newFileRun(h0 *RunList, i, j int) interface{} {
file := h0.At(i).(Spot).File;
// reduce the list of Spots into a list of KindRuns
var h1 RunList;
h1.Vector.Init(j - i);
k := 0;
for ; i < j; i++ {
h1.Set(k, h0.At(i).(Spot).Info);
k++;
}
h2 := h1.reduce(lessKind, newKindRun);
// create the FileRun
groups := make([]*KindRun, h2.Len());
for i := 0; i < h2.Len(); i++ {
groups[i] = h2.At(i).(*KindRun)
}
return &FileRun{file, groups};
}
// ----------------------------------------------------------------------------
// PakRun
// A PakRun describes a run of *FileRuns of a package.
type PakRun struct {
Pak Pak;
Files []*FileRun;
}
// Sorting support for files within a PakRun.
func (p *PakRun) Len() int { return len(p.Files) }
func (p *PakRun) Less(i, j int) bool { return p.Files[i].File.Path < p.Files[j].File.Path }
func (p *PakRun) Swap(i, j int) { p.Files[i], p.Files[j] = p.Files[j], p.Files[i] }
// FileRuns are sorted by package for the reduction into PakRuns.
func lessFileRun(x, y interface{}) bool {
return x.(*FileRun).File.Pak.less(&y.(*FileRun).File.Pak)
}
// newPakRun allocates a new PakRun from the *FileRun run [i, j) in h.
func newPakRun(h *RunList, i, j int) interface{} {
pak := h.At(i).(*FileRun).File.Pak;
files := make([]*FileRun, j-i);
k := 0;
for ; i < j; i++ {
files[k] = h.At(i).(*FileRun);
k++;
}
run := &PakRun{pak, files};
sort.Sort(run); // files were sorted by package; sort them by file now
return run;
}
// ----------------------------------------------------------------------------
// HitList
// A HitList describes a list of PakRuns.
type HitList []*PakRun
// PakRuns are sorted by package.
func lessPakRun(x, y interface{}) bool { return x.(*PakRun).Pak.less(&y.(*PakRun).Pak) }
func reduce(h0 *RunList) HitList {
// reduce a list of Spots into a list of FileRuns
h1 := h0.reduce(lessSpot, newFileRun);
// reduce a list of FileRuns into a list of PakRuns
h2 := h1.reduce(lessFileRun, newPakRun);
// sort the list of PakRuns by package
h2.sort(lessPakRun);
// create a HitList
h := make(HitList, h2.Len());
for i := 0; i < h2.Len(); i++ {
h[i] = h2.At(i).(*PakRun)
}
return h;
}
func (h HitList) filter(pakname string) HitList {
// determine number of matching packages (most of the time just one)
n := 0;
for _, p := range h {
if p.Pak.Name == pakname {
n++
}
}
// create filtered HitList
hh := make(HitList, n);
i := 0;
for _, p := range h {
if p.Pak.Name == pakname {
hh[i] = p;
i++;
}
}
return hh;
}
// ----------------------------------------------------------------------------
// AltWords
type wordPair struct {
canon string; // canonical word spelling (all lowercase)
alt string; // alternative spelling
}
// An AltWords describes a list of alternative spellings for a
// canonical (all lowercase) spelling of a word.
type AltWords struct {
Canon string; // canonical word spelling (all lowercase)
Alts []string; // alternative spelling for the same word
}
// wordPairs are sorted by their canonical spelling.
func lessWordPair(x, y interface{}) bool { return x.(*wordPair).canon < y.(*wordPair).canon }
// newAltWords allocates a new AltWords from the *wordPair run [i, j) in h.
func newAltWords(h *RunList, i, j int) interface{} {
canon := h.At(i).(*wordPair).canon;
alts := make([]string, j-i);
k := 0;
for ; i < j; i++ {
alts[k] = h.At(i).(*wordPair).alt;
k++;
}
return &AltWords{canon, alts};
}
func (a *AltWords) filter(s string) *AltWords {
if len(a.Alts) == 1 && a.Alts[0] == s {
// there are no different alternatives
return nil
}
// make a new AltWords with the current spelling removed
alts := make([]string, len(a.Alts));
i := 0;
for _, w := range a.Alts {
if
|
{ return SpotKind(x >> 1 & 7) }
|
identifier_body
|
index.go
|
// ----------------------------------------------------------------------------
// RunList
// A RunList is a vector of entries that can be sorted according to some
// criteria. A RunList may be compressed by grouping "runs" of entries
// which are equal (according to the sort critera) into a new RunList of
// runs. For instance, a RunList containing pairs (x, y) may be compressed
// into a RunList containing pair runs (x, {y}) where each run consists of
// a list of y's with the same x.
type RunList struct {
vector.Vector;
less func(x, y interface{}) bool;
}
func (h *RunList) Less(i, j int) bool { return h.less(h.At(i), h.At(j)) }
func (h *RunList) sort(less func(x, y interface{}) bool) {
h.less = less;
sort.Sort(h);
}
// Compress entries which are the same according to a sort criteria
// (specified by less) into "runs".
func (h *RunList) reduce(less func(x, y interface{}) bool, newRun func(h *RunList, i, j int) interface{}) *RunList {
// create runs of entries with equal values
h.sort(less);
// for each run, make a new run object and collect them in a new RunList
var hh RunList;
i := 0;
for j := 0; j < h.Len(); j++ {
if less(h.At(i), h.At(j)) {
hh.Push(newRun(h, i, j));
i = j; // start a new run
}
}
// add final run, if any
if i < h.Len() {
hh.Push(newRun(h, i, h.Len()))
}
return &hh;
}
// ----------------------------------------------------------------------------
// SpotInfo
// A SpotInfo value describes a particular identifier spot in a given file;
// It encodes three values: the SpotKind (declaration or use), a line or
// snippet index "lori", and whether it's a line or index.
//
// The following encoding is used:
//
// bits 32 4 1 0
// value [lori|kind|isIndex]
//
type SpotInfo uint32
// SpotKind describes whether an identifier is declared (and what kind of
// declaration) or used.
type SpotKind uint32
const (
PackageClause SpotKind = iota;
ImportDecl;
ConstDecl;
TypeDecl;
VarDecl;
FuncDecl;
MethodDecl;
Use;
nKinds;
)
func init() {
// sanity check: if nKinds is too large, the SpotInfo
// accessor functions may need to be updated
if nKinds > 8 {
panic()
}
}
// makeSpotInfo makes a SpotInfo.
func makeSpotInfo(kind SpotKind, lori int, isIndex bool) SpotInfo {
// encode lori: bits [4..32)
x := SpotInfo(lori) << 4;
if int(x>>4) != lori {
// lori value doesn't fit - since snippet indices are
// most certainly always smaller then 1<<28, this can
// only happen for line numbers; give it no line number (= 0)
x = 0
}
// encode kind: bits [1..4)
x |= SpotInfo(kind) << 1;
// encode isIndex: bit 0
if isIndex {
x |= 1
}
return x;
}
func (x SpotInfo) Kind() SpotKind { return SpotKind(x >> 1 & 7) }
func (x SpotInfo) Lori() int { return int(x >> 4) }
func (x SpotInfo) IsIndex() bool { return x&1 != 0 }
// ----------------------------------------------------------------------------
// KindRun
// Debugging support. Disable to see multiple entries per line.
const removeDuplicates = true
// A KindRun is a run of SpotInfos of the same kind in a given file.
type KindRun struct {
Kind SpotKind;
Infos []SpotInfo;
}
// KindRuns are sorted by line number or index. Since the isIndex bit
// is always the same for all infos in one list we can compare lori's.
func (f *KindRun) Len() int { return len(f.Infos) }
func (f *KindRun) Less(i, j int) bool { return f.Infos[i].Lori() < f.Infos[j].Lori() }
func (f *KindRun) Swap(i, j int) { f.Infos[i], f.Infos[j] = f.Infos[j], f.Infos[i] }
// FileRun contents are sorted by Kind for the reduction into KindRuns.
func lessKind(x, y interface{}) bool { return x.(SpotInfo).Kind() < y.(SpotInfo).Kind() }
// newKindRun allocates a new KindRun from the SpotInfo run [i, j) in h.
func newKindRun(h *RunList, i, j int) interface{} {
kind := h.At(i).(SpotInfo).Kind();
infos := make([]SpotInfo, j-i);
k := 0;
for ; i < j; i++ {
infos[k] = h.At(i).(SpotInfo);
k++;
}
run := &KindRun{kind, infos};
// Spots were sorted by file and kind to create this run.
// Within this run, sort them by line number or index.
sort.Sort(run);
if removeDuplicates {
// Since both the lori and kind field must be
// same for duplicates, and since the isIndex
// bit is always the same for all infos in one
// list we can simply compare the entire info.
k := 0;
var prev SpotInfo;
for i, x := range infos {
if x != prev || i == 0 {
infos[k] = x;
k++;
prev = x;
}
}
run.Infos = infos[0:k];
}
return run;
}
// ----------------------------------------------------------------------------
// FileRun
// A Pak describes a Go package.
type Pak struct {
Path string; // path of directory containing the package
Name string; // package name as declared by package clause
}
// Paks are sorted by name (primary key) and by import path (secondary key).
func (p *Pak) less(q *Pak) bool {
return p.Name < q.Name || p.Name == q.Name && p.Path < q.Path
}
// A File describes a Go file.
type File struct {
Path string; // complete file name
Pak Pak; // the package to which the file belongs
}
// A Spot describes a single occurence of a word.
type Spot struct {
File *File;
Info SpotInfo;
}
// A FileRun is a list of KindRuns belonging to the same file.
type FileRun struct {
File *File;
Groups []*KindRun;
}
// Spots are sorted by path for the reduction into FileRuns.
func lessSpot(x, y interface{}) bool { return x.(Spot).File.Path < y.(Spot).File.Path }
// newFileRun allocates a new FileRun from the Spot run [i, j) in h.
func newFileRun(h0 *RunList, i, j int) interface{} {
file := h0.At(i).(Spot).File;
// reduce the list of Spots into a list of KindRuns
var h1 RunList;
h1.Vector.Init(j - i);
k := 0;
for ; i < j; i++ {
h1.Set(k, h0.At(i).(Spot).Info);
k++;
}
h2 := h1.reduce(lessKind, newKindRun);
// create the FileRun
groups := make([]*KindRun, h2.Len());
for i := 0; i < h2.Len(); i++ {
groups[i] = h2.At(i).(*KindRun)
}
return &FileRun{file, groups};
}
// ----------------------------------------------------------------------------
// PakRun
// A PakRun describes a run of *FileRuns of a package.
type PakRun struct {
Pak Pak;
Files []*FileRun;
}
// Sorting support for files within a PakRun.
func (p *PakRun) Len() int { return len(p.Files) }
func (p *PakRun) Less(i, j int) bool { return p.Files[i].File.Path < p.Files[j].File.Path }
func (p *PakRun) Swap(i, j int) { p.Files[i], p.Files[j] = p.Files[j], p.Files[i] }
// FileRuns are sorted by package for the reduction into PakRuns.
func lessFileRun(x, y interface{}) bool {
return x.(*FileRun).File.Pak.less(&y.(*FileRun).File.Pak)
}
// newPakRun allocates a new PakRun from the *FileRun run [i, j) in h.
func newPakRun(h *RunList, i, j int) interface{} {
pak := h.At(i).(*FileRun).File.Pak;
files := make([]*FileRun, j-i);
k := 0;
for ; i < j; i++ {
files[k] = h.At(i).(*FileRun);
k++;
}
run := &PakRun{
|
)
|
random_line_split
|
|
index.go
|
can
// only happen for line numbers; give it no line number (= 0)
x = 0
}
// encode kind: bits [1..4)
x |= SpotInfo(kind) << 1;
// encode isIndex: bit 0
if isIndex {
x |= 1
}
return x;
}
func (x SpotInfo) Kind() SpotKind { return SpotKind(x >> 1 & 7) }
func (x SpotInfo) Lori() int { return int(x >> 4) }
func (x SpotInfo) IsIndex() bool { return x&1 != 0 }
// ----------------------------------------------------------------------------
// KindRun
// Debugging support. Disable to see multiple entries per line.
const removeDuplicates = true
// A KindRun is a run of SpotInfos of the same kind in a given file.
type KindRun struct {
Kind SpotKind;
Infos []SpotInfo;
}
// KindRuns are sorted by line number or index. Since the isIndex bit
// is always the same for all infos in one list we can compare lori's.
func (f *KindRun) Len() int { return len(f.Infos) }
func (f *KindRun) Less(i, j int) bool { return f.Infos[i].Lori() < f.Infos[j].Lori() }
func (f *KindRun) Swap(i, j int) { f.Infos[i], f.Infos[j] = f.Infos[j], f.Infos[i] }
// FileRun contents are sorted by Kind for the reduction into KindRuns.
func lessKind(x, y interface{}) bool { return x.(SpotInfo).Kind() < y.(SpotInfo).Kind() }
// newKindRun allocates a new KindRun from the SpotInfo run [i, j) in h.
func newKindRun(h *RunList, i, j int) interface{} {
kind := h.At(i).(SpotInfo).Kind();
infos := make([]SpotInfo, j-i);
k := 0;
for ; i < j; i++ {
infos[k] = h.At(i).(SpotInfo);
k++;
}
run := &KindRun{kind, infos};
// Spots were sorted by file and kind to create this run.
// Within this run, sort them by line number or index.
sort.Sort(run);
if removeDuplicates {
// Since both the lori and kind field must be
// same for duplicates, and since the isIndex
// bit is always the same for all infos in one
// list we can simply compare the entire info.
k := 0;
var prev SpotInfo;
for i, x := range infos {
if x != prev || i == 0 {
infos[k] = x;
k++;
prev = x;
}
}
run.Infos = infos[0:k];
}
return run;
}
// ----------------------------------------------------------------------------
// FileRun
// A Pak describes a Go package.
type Pak struct {
Path string; // path of directory containing the package
Name string; // package name as declared by package clause
}
// Paks are sorted by name (primary key) and by import path (secondary key).
func (p *Pak) less(q *Pak) bool {
return p.Name < q.Name || p.Name == q.Name && p.Path < q.Path
}
// A File describes a Go file.
type File struct {
Path string; // complete file name
Pak Pak; // the package to which the file belongs
}
// A Spot describes a single occurence of a word.
type Spot struct {
File *File;
Info SpotInfo;
}
// A FileRun is a list of KindRuns belonging to the same file.
type FileRun struct {
File *File;
Groups []*KindRun;
}
// Spots are sorted by path for the reduction into FileRuns.
func lessSpot(x, y interface{}) bool { return x.(Spot).File.Path < y.(Spot).File.Path }
// newFileRun allocates a new FileRun from the Spot run [i, j) in h.
func newFileRun(h0 *RunList, i, j int) interface{} {
file := h0.At(i).(Spot).File;
// reduce the list of Spots into a list of KindRuns
var h1 RunList;
h1.Vector.Init(j - i);
k := 0;
for ; i < j; i++ {
h1.Set(k, h0.At(i).(Spot).Info);
k++;
}
h2 := h1.reduce(lessKind, newKindRun);
// create the FileRun
groups := make([]*KindRun, h2.Len());
for i := 0; i < h2.Len(); i++ {
groups[i] = h2.At(i).(*KindRun)
}
return &FileRun{file, groups};
}
// ----------------------------------------------------------------------------
// PakRun
// A PakRun describes a run of *FileRuns of a package.
type PakRun struct {
Pak Pak;
Files []*FileRun;
}
// Sorting support for files within a PakRun.
func (p *PakRun) Len() int { return len(p.Files) }
func (p *PakRun) Less(i, j int) bool { return p.Files[i].File.Path < p.Files[j].File.Path }
func (p *PakRun) Swap(i, j int) { p.Files[i], p.Files[j] = p.Files[j], p.Files[i] }
// FileRuns are sorted by package for the reduction into PakRuns.
func lessFileRun(x, y interface{}) bool {
return x.(*FileRun).File.Pak.less(&y.(*FileRun).File.Pak)
}
// newPakRun allocates a new PakRun from the *FileRun run [i, j) in h.
func newPakRun(h *RunList, i, j int) interface{} {
pak := h.At(i).(*FileRun).File.Pak;
files := make([]*FileRun, j-i);
k := 0;
for ; i < j; i++ {
files[k] = h.At(i).(*FileRun);
k++;
}
run := &PakRun{pak, files};
sort.Sort(run); // files were sorted by package; sort them by file now
return run;
}
// ----------------------------------------------------------------------------
// HitList
// A HitList describes a list of PakRuns.
type HitList []*PakRun
// PakRuns are sorted by package.
func lessPakRun(x, y interface{}) bool { return x.(*PakRun).Pak.less(&y.(*PakRun).Pak) }
func reduce(h0 *RunList) HitList {
// reduce a list of Spots into a list of FileRuns
h1 := h0.reduce(lessSpot, newFileRun);
// reduce a list of FileRuns into a list of PakRuns
h2 := h1.reduce(lessFileRun, newPakRun);
// sort the list of PakRuns by package
h2.sort(lessPakRun);
// create a HitList
h := make(HitList, h2.Len());
for i := 0; i < h2.Len(); i++ {
h[i] = h2.At(i).(*PakRun)
}
return h;
}
func (h HitList) filter(pakname string) HitList {
// determine number of matching packages (most of the time just one)
n := 0;
for _, p := range h {
if p.Pak.Name == pakname {
n++
}
}
// create filtered HitList
hh := make(HitList, n);
i := 0;
for _, p := range h {
if p.Pak.Name == pakname {
hh[i] = p;
i++;
}
}
return hh;
}
// ----------------------------------------------------------------------------
// AltWords
type wordPair struct {
canon string; // canonical word spelling (all lowercase)
alt string; // alternative spelling
}
// An AltWords describes a list of alternative spellings for a
// canonical (all lowercase) spelling of a word.
type AltWords struct {
Canon string; // canonical word spelling (all lowercase)
Alts []string; // alternative spelling for the same word
}
// wordPairs are sorted by their canonical spelling.
func lessWordPair(x, y interface{}) bool { return x.(*wordPair).canon < y.(*wordPair).canon }
// newAltWords allocates a new AltWords from the *wordPair run [i, j) in h.
func newAltWords(h *RunList, i, j int) interface{} {
canon := h.At(i).(*wordPair).canon;
alts := make([]string, j-i);
k := 0;
for ; i < j; i++ {
alts[k] = h.At(i).(*wordPair).alt;
k++;
}
return &AltWords{canon, alts};
}
func (a *AltWords)
|
(s string) *AltWords {
if len(a.Alts) == 1 && a.Alts[0] == s {
// there are no different alternatives
return nil
}
// make a new AltWords with the current spelling removed
alts := make([]string, len(a.Alts));
i := 0;
for _, w := range a.Alts {
if
|
filter
|
identifier_name
|
allocator.rs
|
is an assumption that each worker performs the same channel allocation logic; things go wrong otherwise.
pub trait Communicator: 'static {
fn index(&self) -> u64; // number out of peers
fn peers(&self) -> u64; // number of peers
fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>);
}
// TODO : Would be nice if Communicator had associated types for its Pushable and Pullable types,
// TODO : but they would have to be generic over T, with the current set-up. Might require HKT?
// impl<'a, C: Communicator + 'a> Communicator for &'a mut C {
// fn index(&self) -> u64 { (**self).index() }
// fn peers(&self) -> u64 { (**self).peers() }
// fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) { (**self).new_channel() }
// }
// The simplest communicator remains worker-local and just queues sent messages.
pub struct ThreadCommunicator;
impl Communicator for ThreadCommunicator {
fn index(&self) -> u64 { 0 }
fn
|
(&self) -> u64 { 1 }
fn new_channel<T:'static>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let shared = Rc::new(RefCell::new(VecDeque::<T>::new()));
return (vec![Box::new(shared.clone()) as Box<Pushable<T>>], Box::new(shared.clone()) as Box<Pullable<T>>)
}
}
// A specific Communicator for inter-thread intra-process communication
pub struct ProcessCommunicator {
inner: ThreadCommunicator, // inner ThreadCommunicator
index: u64, // number out of peers
peers: u64, // number of peer allocators (for typed channel allocation).
allocated: u64, // indicates how many have been allocated (locally).
channels: Arc<Mutex<Vec<Box<Any+Send>>>>, // Box<Any+Send> -> Box<Vec<Option<(Vec<Sender<T>>, Receiver<T>)>>>
}
impl ProcessCommunicator {
pub fn inner<'a>(&'a mut self) -> &'a mut ThreadCommunicator { &mut self.inner }
pub fn new_vector(count: u64) -> Vec<ProcessCommunicator> {
let channels = Arc::new(Mutex::new(Vec::new()));
return (0 .. count).map(|index| ProcessCommunicator {
inner: ThreadCommunicator,
index: index,
peers: count,
allocated: 0,
channels: channels.clone(),
}).collect();
}
}
impl Communicator for ProcessCommunicator {
fn index(&self) -> u64 { self.index }
fn peers(&self) -> u64 { self.peers }
fn new_channel<T:Send+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let mut channels = self.channels.lock().ok().expect("mutex error?");
if self.allocated == channels.len() as u64 { // we need a new channel ...
let mut senders = Vec::new();
let mut receivers = Vec::new();
for _ in (0..self.peers) {
let (s, r): (Sender<T>, Receiver<T>) = channel();
senders.push(s);
receivers.push(r);
}
let mut to_box = Vec::new();
for recv in receivers.drain_temp() {
to_box.push(Some((senders.clone(), recv)));
}
channels.push(Box::new(to_box));
}
match channels[self.allocated as usize].downcast_mut::<(Vec<Option<(Vec<Sender<T>>, Receiver<T>)>>)>() {
Some(ref mut vector) => {
self.allocated += 1;
let (mut send, recv) = vector[self.index as usize].take().unwrap();
let mut temp = Vec::new();
for s in send.drain_temp() { temp.push(Box::new(s) as Box<Pushable<T>>); }
return (temp, Box::new(recv) as Box<Pullable<T>>)
}
_ => { panic!("unable to cast channel correctly"); }
}
}
}
// A communicator intended for binary channels (networking, pipes, shared memory)
pub struct BinaryCommunicator {
pub inner: ProcessCommunicator, // inner ProcessCommunicator (use for process-local channels)
pub index: u64, // index of this worker
pub peers: u64, // number of peer workers
pub graph: u64, // identifier for the current graph
pub allocated: u64, // indicates how many channels have been allocated (locally).
// for loading up state in the networking threads.
pub writers: Vec<Sender<((u64, u64, u64), Sender<Vec<u8>>)>>, // (index, back-to-worker)
pub readers: Vec<Sender<((u64, u64, u64), Sender<Vec<u8>>, Receiver<Vec<u8>>)>>, // (index, data-to-worker, back-from-worker)
pub senders: Vec<Sender<(MessageHeader, Vec<u8>)>> // for sending bytes!
}
impl BinaryCommunicator {
pub fn inner<'a>(&'a mut self) -> &'a mut ProcessCommunicator { &mut self.inner }
}
// A Communicator backed by Sender<Vec<u8>>/Receiver<Vec<u8>> pairs (e.g. networking, shared memory, files, pipes)
impl Communicator for BinaryCommunicator {
fn index(&self) -> u64 { self.index }
fn peers(&self) -> u64 { self.peers }
fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let mut pushers: Vec<Box<Pushable<T>>> = Vec::new(); // built-up vector of Box<Pushable<T>> to return
// we'll need process-local channels as well (no self-loop binary connection in this design; perhaps should allow)
let inner_peers = self.inner.peers();
let (inner_sends, inner_recv) = self.inner.new_channel();
// prep a pushable for each endpoint, multiplied by inner_peers
for (index, writer) in self.writers.iter().enumerate() {
for _ in (0..inner_peers) {
let (s,r) = channel(); // generate a binary (Vec<u8>) channel pair of (back_to_worker, back_from_net)
let target_index = if index as u64 >= (self.index * inner_peers) { index as u64 + inner_peers } else { index as u64 };
println!("init'ing send channel: ({} {} {})", self.index, self.graph, self.allocated);
writer.send(((self.index, self.graph, self.allocated), s)).unwrap();
let header = MessageHeader {
graph: self.graph,
channel: self.allocated,
source: self.index,
target: target_index,
length: 0,
};
pushers.push(Box::new(BinaryPushable::new(header, self.senders[index].clone(), r)));
}
}
// splice inner_sends into the vector of pushables
for (index, writer) in inner_sends.into_iter().enumerate() {
pushers.insert((self.index * inner_peers) as usize + index, writer);
}
// prep a Box<Pullable<T>> using inner_recv and fresh registered pullables
let (send,recv) = channel(); // binary channel from binary listener to BinaryPullable<T>
let mut pullsends = Vec::new();
for reader in self.readers.iter() {
let (s,r) = channel();
pullsends.push(s);
println!("init'ing recv channel: ({} {} {})", self.index, self.graph, self.allocated);
reader.send(((self.index, self.graph, self.allocated), send.clone(), r)).unwrap();
}
let pullable = Box::new(BinaryPullable {
inner: inner_recv,
senders: pullsends,
receiver: recv,
stack: Default::default(),
});
self.allocated += 1;
return (pushers, pullable);
}
}
struct BinaryPushable<T: Columnar> {
header: MessageHeader,
sender: Sender<(MessageHeader, Vec<u8>)>, // targets for each remote destination
receiver: Receiver<Vec<u8>>, // source of empty binary vectors
phantom: PhantomData<T>,
buffer: Vec<u8>,
stack: <T as Columnar>::Stack,
}
impl<T: Columnar> BinaryPushable<T> {
pub fn new(header: MessageHeader, sender: Sender<(MessageHeader, Vec<u8>)>, receiver: Receiver<Vec<u8>>) -> BinaryPushable<T> {
BinaryPushable {
header: header,
sender: sender,
receiver: receiver,
phantom:
|
peers
|
identifier_name
|
allocator.rs
|
pub trait Communicator: 'static {
fn index(&self) -> u64; // number out of peers
fn peers(&self) -> u64; // number of peers
fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>);
}
// TODO : Would be nice if Communicator had associated types for its Pushable and Pullable types,
// TODO : but they would have to be generic over T, with the current set-up. Might require HKT?
// impl<'a, C: Communicator + 'a> Communicator for &'a mut C {
// fn index(&self) -> u64 { (**self).index() }
// fn peers(&self) -> u64 { (**self).peers() }
// fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) { (**self).new_channel() }
// }
// The simplest communicator remains worker-local and just queues sent messages.
pub struct ThreadCommunicator;
impl Communicator for ThreadCommunicator {
fn index(&self) -> u64 { 0 }
fn peers(&self) -> u64 { 1 }
fn new_channel<T:'static>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let shared = Rc::new(RefCell::new(VecDeque::<T>::new()));
return (vec![Box::new(shared.clone()) as Box<Pushable<T>>], Box::new(shared.clone()) as Box<Pullable<T>>)
}
}
// A specific Communicator for inter-thread intra-process communication
pub struct ProcessCommunicator {
inner: ThreadCommunicator, // inner ThreadCommunicator
index: u64, // number out of peers
peers: u64, // number of peer allocators (for typed channel allocation).
allocated: u64, // indicates how many have been allocated (locally).
channels: Arc<Mutex<Vec<Box<Any+Send>>>>, // Box<Any+Send> -> Box<Vec<Option<(Vec<Sender<T>>, Receiver<T>)>>>
}
impl ProcessCommunicator {
pub fn inner<'a>(&'a mut self) -> &'a mut ThreadCommunicator { &mut self.inner }
pub fn new_vector(count: u64) -> Vec<ProcessCommunicator> {
let channels = Arc::new(Mutex::new(Vec::new()));
return (0 .. count).map(|index| ProcessCommunicator {
inner: ThreadCommunicator,
index: index,
peers: count,
allocated: 0,
channels: channels.clone(),
}).collect();
}
}
impl Communicator for ProcessCommunicator {
fn index(&self) -> u64 { self.index }
fn peers(&self) -> u64 { self.peers }
fn new_channel<T:Send+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let mut channels = self.channels.lock().ok().expect("mutex error?");
if self.allocated == channels.len() as u64 { // we need a new channel ...
let mut senders = Vec::new();
let mut receivers = Vec::new();
for _ in (0..self.peers) {
let (s, r): (Sender<T>, Receiver<T>) = channel();
senders.push(s);
receivers.push(r);
}
let mut to_box = Vec::new();
for recv in receivers.drain_temp() {
to_box.push(Some((senders.clone(), recv)));
}
channels.push(Box::new(to_box));
}
match channels[self.allocated as usize].downcast_mut::<(Vec<Option<(Vec<Sender<T>>, Receiver<T>)>>)>() {
Some(ref mut vector) => {
self.allocated += 1;
let (mut send, recv) = vector[self.index as usize].take().unwrap();
let mut temp = Vec::new();
for s in send.drain_temp() { temp.push(Box::new(s) as Box<Pushable<T>>); }
return (temp, Box::new(recv) as Box<Pullable<T>>)
}
_ => { panic!("unable to cast channel correctly"); }
}
}
}
// A communicator intended for binary channels (networking, pipes, shared memory)
pub struct BinaryCommunicator {
pub inner: ProcessCommunicator, // inner ProcessCommunicator (use for process-local channels)
pub index: u64, // index of this worker
pub peers: u64, // number of peer workers
pub graph: u64, // identifier for the current graph
pub allocated: u64, // indicates how many channels have been allocated (locally).
// for loading up state in the networking threads.
pub writers: Vec<Sender<((u64, u64, u64), Sender<Vec<u8>>)>>, // (index, back-to-worker)
pub readers: Vec<Sender<((u64, u64, u64), Sender<Vec<u8>>, Receiver<Vec<u8>>)>>, // (index, data-to-worker, back-from-worker)
pub senders: Vec<Sender<(MessageHeader, Vec<u8>)>> // for sending bytes!
}
impl BinaryCommunicator {
pub fn inner<'a>(&'a mut self) -> &'a mut ProcessCommunicator { &mut self.inner }
}
// A Communicator backed by Sender<Vec<u8>>/Receiver<Vec<u8>> pairs (e.g. networking, shared memory, files, pipes)
impl Communicator for BinaryCommunicator {
fn index(&self) -> u64 { self.index }
fn peers(&self) -> u64 { self.peers }
fn new_channel<T:Send+Columnar+Any>(&mut self) -> (Vec<Box<Pushable<T>>>, Box<Pullable<T>>) {
let mut pushers: Vec<Box<Pushable<T>>> = Vec::new(); // built-up vector of Box<Pushable<T>> to return
// we'll need process-local channels as well (no self-loop binary connection in this design; perhaps should allow)
let inner_peers = self.inner.peers();
let (inner_sends, inner_recv) = self.inner.new_channel();
// prep a pushable for each endpoint, multiplied by inner_peers
for (index, writer) in self.writers.iter().enumerate() {
for _ in (0..inner_peers) {
let (s,r) = channel(); // generate a binary (Vec<u8>) channel pair of (back_to_worker, back_from_net)
let target_index = if index as u64 >= (self.index * inner_peers) { index as u64 + inner_peers } else { index as u64 };
println!("init'ing send channel: ({} {} {})", self.index, self.graph, self.allocated);
writer.send(((self.index, self.graph, self.allocated), s)).unwrap();
let header = MessageHeader {
graph: self.graph,
channel: self.allocated,
source: self.index,
target: target_index,
length: 0,
};
pushers.push(Box::new(BinaryPushable::new(header, self.senders[index].clone(), r)));
}
}
// splice inner_sends into the vector of pushables
for (index, writer) in inner_sends.into_iter().enumerate() {
pushers.insert((self.index * inner_peers) as usize + index, writer);
}
// prep a Box<Pullable<T>> using inner_recv and fresh registered pullables
let (send,recv) = channel(); // binary channel from binary listener to BinaryPullable<T>
let mut pullsends = Vec::new();
for reader in self.readers.iter() {
let (s,r) = channel();
pullsends.push(s);
println!("init'ing recv channel: ({} {} {})", self.index, self.graph, self.allocated);
reader.send(((self.index, self.graph, self.allocated), send.clone(), r)).unwrap();
}
let pullable = Box::new(BinaryPullable {
inner: inner_recv,
senders: pullsends,
receiver: recv,
stack: Default::default(),
});
self.allocated += 1;
return (pushers, pullable);
}
}
struct BinaryPushable<T: Columnar> {
header: MessageHeader,
sender: Sender<(MessageHeader, Vec<u8>)>, // targets for each remote destination
receiver: Receiver<Vec<u8>>, // source of empty binary vectors
phantom: PhantomData<T>,
buffer: Vec<u8>,
stack: <T as Columnar>::Stack,
}
impl<T: Columnar> BinaryPushable<T> {
pub fn new(header: MessageHeader, sender: Sender<(MessageHeader, Vec<u8>)>, receiver: Receiver<Vec<u8>>) -> BinaryPushable<T> {
BinaryPushable {
header: header,
|
sender: sender,
receiver: receiver,
phantom: PhantomData,
buffer: Vec::new(),
stack: Default::default(),
|
random_line_split
|
|
load_balancer.rs
|
may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
use std::convert::TryFrom;
use std::sync::atomic::{AtomicUsize, Ordering};
use rand::{thread_rng, Rng};
use serde::{Deserialize, Serialize};
use crate::{config::UpstreamEndpoints, filters::prelude::*, map_proto_enum};
crate::include_proto!("quilkin.extensions.filters.load_balancer.v1alpha1");
use self::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::Policy as ProtoPolicy, LoadBalancer as ProtoConfig,
};
/// Policy represents how a [`LoadBalancerFilter`] distributes
/// packets across endpoints.
#[derive(Debug, Deserialize, Serialize, Eq, PartialEq)]
pub enum Policy {
/// Send packets to endpoints in turns.
#[serde(rename = "ROUND_ROBIN")]
RoundRobin,
/// Send packets to endpoints chosen at random.
#[serde(rename = "RANDOM")]
Random,
}
impl Default for Policy {
fn default() -> Self {
Policy::RoundRobin
}
}
/// Config represents configuration for a [`LoadBalancerFilter`].
#[derive(Serialize, Deserialize, Debug, PartialEq)]
struct Config {
#[serde(default)]
policy: Policy,
}
impl TryFrom<ProtoConfig> for Config {
type Error = ConvertProtoConfigError;
fn try_from(p: ProtoConfig) -> Result<Self, Self::Error> {
let policy = p
.policy
.map(|policy| {
map_proto_enum!(
value = policy.value,
field = "policy",
proto_enum_type = ProtoPolicy,
target_enum_type = Policy,
variants = [RoundRobin, Random]
)
})
.transpose()?
.unwrap_or_else(Policy::default);
Ok(Self { policy })
}
}
/// EndpointChooser chooses from a set of endpoints that a proxy is connected to.
trait EndpointChooser: Send + Sync {
/// choose_endpoints asks for the next endpoint(s) to use.
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints);
}
/// RoundRobinEndpointChooser chooses endpoints in round-robin order.
pub struct RoundRobinEndpointChooser {
next_endpoint: AtomicUsize,
}
impl RoundRobinEndpointChooser {
fn new() -> Self {
RoundRobinEndpointChooser {
next_endpoint: AtomicUsize::new(0),
}
}
}
impl EndpointChooser for RoundRobinEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
let count = self.next_endpoint.fetch_add(1, Ordering::Relaxed);
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let num_endpoints = endpoints.size();
endpoints.keep(count % num_endpoints)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// RandomEndpointChooser chooses endpoints in random order.
pub struct RandomEndpointChooser;
impl EndpointChooser for RandomEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let idx = (&mut thread_rng()).gen_range(0..endpoints.size());
endpoints.keep(idx)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// Creates instances of LoadBalancerFilter.
#[derive(Default)]
pub struct LoadBalancerFilterFactory;
/// LoadBalancerFilter load balances packets over the upstream endpoints.
#[crate::filter("quilkin.extensions.filters.load_balancer.v1alpha1.LoadBalancer")]
struct LoadBalancerFilter {
endpoint_chooser: Box<dyn EndpointChooser>,
}
impl FilterFactory for LoadBalancerFilterFactory {
fn name(&self) -> &'static str {
LoadBalancerFilter::FILTER_NAME
}
fn create_filter(&self, args: CreateFilterArgs) -> Result<Box<dyn Filter>, Error> {
let config: Config = self
.require_config(args.config)?
.deserialize::<Config, ProtoConfig>(self.name())?;
let endpoint_chooser: Box<dyn EndpointChooser> = match config.policy {
Policy::RoundRobin => Box::new(RoundRobinEndpointChooser::new()),
Policy::Random => Box::new(RandomEndpointChooser),
};
Ok(Box::new(LoadBalancerFilter { endpoint_chooser }))
}
}
impl Filter for LoadBalancerFilter {
fn read(&self, mut ctx: ReadContext) -> Option<ReadResponse> {
self.endpoint_chooser.choose_endpoints(&mut ctx.endpoints);
Some(ctx.into())
}
}
#[cfg(test)]
mod tests {
use std::collections::HashSet;
use std::convert::TryFrom;
use std::net::SocketAddr;
use super::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::{Policy as ProtoPolicy, PolicyValue},
LoadBalancer as ProtoConfig,
};
use super::{Config, Policy};
use crate::cluster::Endpoint;
use crate::config::Endpoints;
use crate::filters::{
extensions::load_balancer::LoadBalancerFilterFactory, CreateFilterArgs, Filter,
FilterFactory, ReadContext,
};
use prometheus::Registry;
fn create_filter(config: &str) -> Box<dyn Filter> {
let factory = LoadBalancerFilterFactory;
factory
.create_filter(CreateFilterArgs::fixed(
Registry::default(),
Some(&serde_yaml::from_str(config).unwrap()),
))
.unwrap()
}
fn get_response_addresses(
filter: &dyn Filter,
input_addresses: &[SocketAddr],
) -> Vec<SocketAddr> {
filter
.read(ReadContext::new(
Endpoints::new(
input_addresses
.iter()
.map(|addr| Endpoint::from_address(*addr))
.collect(),
)
.unwrap()
.into(),
"127.0.0.1:8080".parse().unwrap(),
vec![],
))
.unwrap()
.endpoints
.iter()
.map(|ep| ep.address)
.collect::<Vec<_>>()
}
#[test]
fn convert_proto_config() {
let test_cases = vec![
(
"RandomPolicy",
ProtoConfig {
policy: Some(PolicyValue {
value: ProtoPolicy::Random as i32,
}),
},
Some(Config {
policy: Policy::Random,
}),
),
(
"RoundRobinPolicy",
ProtoConfig {
policy: Some(PolicyValue {
value: ProtoPolicy::RoundRobin as i32,
}),
},
Some(Config {
policy: Policy::RoundRobin,
}),
),
(
"should fail when invalid policy is provided",
ProtoConfig {
policy: Some(PolicyValue { value: 42 }),
},
None,
),
(
"should use correct default values",
ProtoConfig { policy: None },
Some(Config {
policy: Policy::default(),
}),
),
];
for (name, proto_config, expected) in test_cases {
let result = Config::try_from(proto_config);
assert_eq!(
result.is_err(),
expected.is_none(),
"{}: error expectation does not match",
name
);
if let Some(expected) = expected
|
}
}
#[test]
fn round_robin_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: ROUND_ROBIN
";
let filter = create_filter(yaml);
// Check that we repeat the same addresses in sequence forever.
let expected_sequence = addresses.iter().map(|addr| vec![*addr]).collect::<Vec<_>>();
for _ in 0..10 {
assert_eq!(
expected_sequence,
(0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>()
);
}
}
#[test]
fn random_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: RANDOM
";
let filter = create_filter(yaml);
// Run a few selection rounds through the addresses.
let mut result_sequences = vec![];
for _ in 0..10 {
let sequence = (0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>();
result_sequences.push(sequence);
}
// Check that every address was chosen at least once.
assert_eq
|
{
assert_eq!(expected, result.unwrap(), "{}", name);
}
|
conditional_block
|
load_balancer.rs
|
may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
use std::convert::TryFrom;
use std::sync::atomic::{AtomicUsize, Ordering};
use rand::{thread_rng, Rng};
use serde::{Deserialize, Serialize};
use crate::{config::UpstreamEndpoints, filters::prelude::*, map_proto_enum};
crate::include_proto!("quilkin.extensions.filters.load_balancer.v1alpha1");
use self::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::Policy as ProtoPolicy, LoadBalancer as ProtoConfig,
};
/// Policy represents how a [`LoadBalancerFilter`] distributes
/// packets across endpoints.
#[derive(Debug, Deserialize, Serialize, Eq, PartialEq)]
pub enum Policy {
/// Send packets to endpoints in turns.
#[serde(rename = "ROUND_ROBIN")]
RoundRobin,
/// Send packets to endpoints chosen at random.
#[serde(rename = "RANDOM")]
Random,
}
impl Default for Policy {
fn default() -> Self {
Policy::RoundRobin
}
}
/// Config represents configuration for a [`LoadBalancerFilter`].
#[derive(Serialize, Deserialize, Debug, PartialEq)]
struct
|
{
#[serde(default)]
policy: Policy,
}
impl TryFrom<ProtoConfig> for Config {
type Error = ConvertProtoConfigError;
fn try_from(p: ProtoConfig) -> Result<Self, Self::Error> {
let policy = p
.policy
.map(|policy| {
map_proto_enum!(
value = policy.value,
field = "policy",
proto_enum_type = ProtoPolicy,
target_enum_type = Policy,
variants = [RoundRobin, Random]
)
})
.transpose()?
.unwrap_or_else(Policy::default);
Ok(Self { policy })
}
}
/// EndpointChooser chooses from a set of endpoints that a proxy is connected to.
trait EndpointChooser: Send + Sync {
/// choose_endpoints asks for the next endpoint(s) to use.
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints);
}
/// RoundRobinEndpointChooser chooses endpoints in round-robin order.
pub struct RoundRobinEndpointChooser {
next_endpoint: AtomicUsize,
}
impl RoundRobinEndpointChooser {
fn new() -> Self {
RoundRobinEndpointChooser {
next_endpoint: AtomicUsize::new(0),
}
}
}
impl EndpointChooser for RoundRobinEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
let count = self.next_endpoint.fetch_add(1, Ordering::Relaxed);
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let num_endpoints = endpoints.size();
endpoints.keep(count % num_endpoints)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// RandomEndpointChooser chooses endpoints in random order.
pub struct RandomEndpointChooser;
impl EndpointChooser for RandomEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let idx = (&mut thread_rng()).gen_range(0..endpoints.size());
endpoints.keep(idx)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// Creates instances of LoadBalancerFilter.
#[derive(Default)]
pub struct LoadBalancerFilterFactory;
/// LoadBalancerFilter load balances packets over the upstream endpoints.
#[crate::filter("quilkin.extensions.filters.load_balancer.v1alpha1.LoadBalancer")]
struct LoadBalancerFilter {
endpoint_chooser: Box<dyn EndpointChooser>,
}
impl FilterFactory for LoadBalancerFilterFactory {
fn name(&self) -> &'static str {
LoadBalancerFilter::FILTER_NAME
}
fn create_filter(&self, args: CreateFilterArgs) -> Result<Box<dyn Filter>, Error> {
let config: Config = self
.require_config(args.config)?
.deserialize::<Config, ProtoConfig>(self.name())?;
let endpoint_chooser: Box<dyn EndpointChooser> = match config.policy {
Policy::RoundRobin => Box::new(RoundRobinEndpointChooser::new()),
Policy::Random => Box::new(RandomEndpointChooser),
};
Ok(Box::new(LoadBalancerFilter { endpoint_chooser }))
}
}
impl Filter for LoadBalancerFilter {
fn read(&self, mut ctx: ReadContext) -> Option<ReadResponse> {
self.endpoint_chooser.choose_endpoints(&mut ctx.endpoints);
Some(ctx.into())
}
}
#[cfg(test)]
mod tests {
use std::collections::HashSet;
use std::convert::TryFrom;
use std::net::SocketAddr;
use super::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::{Policy as ProtoPolicy, PolicyValue},
LoadBalancer as ProtoConfig,
};
use super::{Config, Policy};
use crate::cluster::Endpoint;
use crate::config::Endpoints;
use crate::filters::{
extensions::load_balancer::LoadBalancerFilterFactory, CreateFilterArgs, Filter,
FilterFactory, ReadContext,
};
use prometheus::Registry;
fn create_filter(config: &str) -> Box<dyn Filter> {
let factory = LoadBalancerFilterFactory;
factory
.create_filter(CreateFilterArgs::fixed(
Registry::default(),
Some(&serde_yaml::from_str(config).unwrap()),
))
.unwrap()
}
fn get_response_addresses(
filter: &dyn Filter,
input_addresses: &[SocketAddr],
) -> Vec<SocketAddr> {
filter
.read(ReadContext::new(
Endpoints::new(
input_addresses
.iter()
.map(|addr| Endpoint::from_address(*addr))
.collect(),
)
.unwrap()
.into(),
"127.0.0.1:8080".parse().unwrap(),
vec![],
))
.unwrap()
.endpoints
.iter()
.map(|ep| ep.address)
.collect::<Vec<_>>()
}
#[test]
fn convert_proto_config() {
let test_cases = vec![
(
"RandomPolicy",
ProtoConfig {
policy: Some(PolicyValue {
value: ProtoPolicy::Random as i32,
}),
},
Some(Config {
policy: Policy::Random,
}),
),
(
"RoundRobinPolicy",
ProtoConfig {
policy: Some(PolicyValue {
value: ProtoPolicy::RoundRobin as i32,
}),
},
Some(Config {
policy: Policy::RoundRobin,
}),
),
(
"should fail when invalid policy is provided",
ProtoConfig {
policy: Some(PolicyValue { value: 42 }),
},
None,
),
(
"should use correct default values",
ProtoConfig { policy: None },
Some(Config {
policy: Policy::default(),
}),
),
];
for (name, proto_config, expected) in test_cases {
let result = Config::try_from(proto_config);
assert_eq!(
result.is_err(),
expected.is_none(),
"{}: error expectation does not match",
name
);
if let Some(expected) = expected {
assert_eq!(expected, result.unwrap(), "{}", name);
}
}
}
#[test]
fn round_robin_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: ROUND_ROBIN
";
let filter = create_filter(yaml);
// Check that we repeat the same addresses in sequence forever.
let expected_sequence = addresses.iter().map(|addr| vec![*addr]).collect::<Vec<_>>();
for _ in 0..10 {
assert_eq!(
expected_sequence,
(0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>()
);
}
}
#[test]
fn random_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: RANDOM
";
let filter = create_filter(yaml);
// Run a few selection rounds through the addresses.
let mut result_sequences = vec![];
for _ in 0..10 {
let sequence = (0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>();
result_sequences.push(sequence);
}
// Check that every address was chosen at least once.
assert_eq!(
|
Config
|
identifier_name
|
load_balancer.rs
|
You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
use std::convert::TryFrom;
use std::sync::atomic::{AtomicUsize, Ordering};
use rand::{thread_rng, Rng};
use serde::{Deserialize, Serialize};
use crate::{config::UpstreamEndpoints, filters::prelude::*, map_proto_enum};
crate::include_proto!("quilkin.extensions.filters.load_balancer.v1alpha1");
use self::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::Policy as ProtoPolicy, LoadBalancer as ProtoConfig,
};
/// Policy represents how a [`LoadBalancerFilter`] distributes
/// packets across endpoints.
#[derive(Debug, Deserialize, Serialize, Eq, PartialEq)]
pub enum Policy {
/// Send packets to endpoints in turns.
#[serde(rename = "ROUND_ROBIN")]
RoundRobin,
/// Send packets to endpoints chosen at random.
#[serde(rename = "RANDOM")]
Random,
}
impl Default for Policy {
fn default() -> Self {
Policy::RoundRobin
}
}
/// Config represents configuration for a [`LoadBalancerFilter`].
#[derive(Serialize, Deserialize, Debug, PartialEq)]
struct Config {
#[serde(default)]
policy: Policy,
}
impl TryFrom<ProtoConfig> for Config {
type Error = ConvertProtoConfigError;
fn try_from(p: ProtoConfig) -> Result<Self, Self::Error> {
let policy = p
.policy
.map(|policy| {
map_proto_enum!(
value = policy.value,
field = "policy",
proto_enum_type = ProtoPolicy,
target_enum_type = Policy,
variants = [RoundRobin, Random]
)
})
.transpose()?
.unwrap_or_else(Policy::default);
Ok(Self { policy })
}
}
/// EndpointChooser chooses from a set of endpoints that a proxy is connected to.
trait EndpointChooser: Send + Sync {
/// choose_endpoints asks for the next endpoint(s) to use.
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints);
}
/// RoundRobinEndpointChooser chooses endpoints in round-robin order.
pub struct RoundRobinEndpointChooser {
next_endpoint: AtomicUsize,
}
impl RoundRobinEndpointChooser {
fn new() -> Self {
RoundRobinEndpointChooser {
next_endpoint: AtomicUsize::new(0),
}
}
}
impl EndpointChooser for RoundRobinEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
let count = self.next_endpoint.fetch_add(1, Ordering::Relaxed);
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let num_endpoints = endpoints.size();
endpoints.keep(count % num_endpoints)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// RandomEndpointChooser chooses endpoints in random order.
pub struct RandomEndpointChooser;
impl EndpointChooser for RandomEndpointChooser {
fn choose_endpoints(&self, endpoints: &mut UpstreamEndpoints) {
// Note: Unwrap is safe here because the index is guaranteed to be in range.
let idx = (&mut thread_rng()).gen_range(0..endpoints.size());
endpoints.keep(idx)
.expect("BUG: unwrap should have been safe because index into endpoints list should be in range");
}
}
/// Creates instances of LoadBalancerFilter.
#[derive(Default)]
pub struct LoadBalancerFilterFactory;
/// LoadBalancerFilter load balances packets over the upstream endpoints.
#[crate::filter("quilkin.extensions.filters.load_balancer.v1alpha1.LoadBalancer")]
struct LoadBalancerFilter {
endpoint_chooser: Box<dyn EndpointChooser>,
}
impl FilterFactory for LoadBalancerFilterFactory {
fn name(&self) -> &'static str {
LoadBalancerFilter::FILTER_NAME
}
fn create_filter(&self, args: CreateFilterArgs) -> Result<Box<dyn Filter>, Error> {
let config: Config = self
.require_config(args.config)?
.deserialize::<Config, ProtoConfig>(self.name())?;
let endpoint_chooser: Box<dyn EndpointChooser> = match config.policy {
Policy::RoundRobin => Box::new(RoundRobinEndpointChooser::new()),
Policy::Random => Box::new(RandomEndpointChooser),
};
Ok(Box::new(LoadBalancerFilter { endpoint_chooser }))
}
}
impl Filter for LoadBalancerFilter {
fn read(&self, mut ctx: ReadContext) -> Option<ReadResponse> {
self.endpoint_chooser.choose_endpoints(&mut ctx.endpoints);
Some(ctx.into())
}
}
#[cfg(test)]
mod tests {
use std::collections::HashSet;
use std::convert::TryFrom;
use std::net::SocketAddr;
use super::quilkin::extensions::filters::load_balancer::v1alpha1::{
load_balancer::{Policy as ProtoPolicy, PolicyValue},
LoadBalancer as ProtoConfig,
};
use super::{Config, Policy};
use crate::cluster::Endpoint;
use crate::config::Endpoints;
use crate::filters::{
extensions::load_balancer::LoadBalancerFilterFactory, CreateFilterArgs, Filter,
FilterFactory, ReadContext,
};
use prometheus::Registry;
fn create_filter(config: &str) -> Box<dyn Filter> {
let factory = LoadBalancerFilterFactory;
factory
.create_filter(CreateFilterArgs::fixed(
Registry::default(),
Some(&serde_yaml::from_str(config).unwrap()),
))
.unwrap()
}
fn get_response_addresses(
filter: &dyn Filter,
input_addresses: &[SocketAddr],
) -> Vec<SocketAddr> {
filter
.read(ReadContext::new(
Endpoints::new(
input_addresses
.iter()
.map(|addr| Endpoint::from_address(*addr))
.collect(),
)
.unwrap()
.into(),
"127.0.0.1:8080".parse().unwrap(),
vec![],
))
.unwrap()
.endpoints
.iter()
.map(|ep| ep.address)
.collect::<Vec<_>>()
}
#[test]
|
policy: Some(PolicyValue {
value: ProtoPolicy::Random as i32,
}),
},
Some(Config {
policy: Policy::Random,
}),
),
(
"RoundRobinPolicy",
ProtoConfig {
policy: Some(PolicyValue {
value: ProtoPolicy::RoundRobin as i32,
}),
},
Some(Config {
policy: Policy::RoundRobin,
}),
),
(
"should fail when invalid policy is provided",
ProtoConfig {
policy: Some(PolicyValue { value: 42 }),
},
None,
),
(
"should use correct default values",
ProtoConfig { policy: None },
Some(Config {
policy: Policy::default(),
}),
),
];
for (name, proto_config, expected) in test_cases {
let result = Config::try_from(proto_config);
assert_eq!(
result.is_err(),
expected.is_none(),
"{}: error expectation does not match",
name
);
if let Some(expected) = expected {
assert_eq!(expected, result.unwrap(), "{}", name);
}
}
}
#[test]
fn round_robin_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: ROUND_ROBIN
";
let filter = create_filter(yaml);
// Check that we repeat the same addresses in sequence forever.
let expected_sequence = addresses.iter().map(|addr| vec![*addr]).collect::<Vec<_>>();
for _ in 0..10 {
assert_eq!(
expected_sequence,
(0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>()
);
}
}
#[test]
fn random_load_balancer_policy() {
let addresses = vec![
"127.0.0.1:8080".parse().unwrap(),
"127.0.0.2:8080".parse().unwrap(),
"127.0.0.3:8080".parse().unwrap(),
];
let yaml = "
policy: RANDOM
";
let filter = create_filter(yaml);
// Run a few selection rounds through the addresses.
let mut result_sequences = vec![];
for _ in 0..10 {
let sequence = (0..addresses.len())
.map(|_| get_response_addresses(filter.as_ref(), &addresses))
.collect::<Vec<_>>();
result_sequences.push(sequence);
}
// Check that every address was chosen at least once.
assert_eq!(
|
fn convert_proto_config() {
let test_cases = vec![
(
"RandomPolicy",
ProtoConfig {
|
random_line_split
|
usher.go
|
qualified domain whose mappings we want
DBPath string // full path to database for Domain
ConfigPath string // full path to usher config file
}
type Entry struct {
Code string
Url string
}
type ConfigEntry struct {
Type string `yaml:"type"`
AWSKey string `yaml:"aws_key,omitempty"`
AWSSecret string `yaml:"aws_secret,omitempty"`
AWSRegion string `yaml:"aws_region,omitempty"`
}
// NewDB creates a DB struct with members derived from parameters,
// the environment, or defaults (in that order). It does no checking
// that the values produced are sane or exist on the filesystem.
func NewDB(domain string) (*DB, error) {
// Get root
root := os.Getenv("USHER_ROOT")
if root == "" {
// If USHER_ROOT is unset, check if there is an usher.yml in the cwd
stat, err := os.Stat("usher.yml")
if err == nil && !stat.IsDir()
|
}
if root == "" {
// If root is still unset, default to "os.UserConfigDir()/usher"
configDir, err := os.UserConfigDir()
if err != nil {
return nil, err
}
root = filepath.Join(configDir, "usher")
}
// Derive domain if not set - check for USHER_DOMAIN in environment
if domain == "" {
domain = os.Getenv("USHER_DOMAIN")
}
// Else infer the domain if only one database exists
if domain == "" {
matches, _ := filepath.Glob(filepath.Join(root, "*.*.yml"))
if len(matches) == 1 {
// Exactly one match - strip .yml suffix to get domain
re := regexp.MustCompile(`.yml$`)
domain = re.ReplaceAllLiteralString(filepath.Base(matches[0]), "")
}
}
// Else give up with an error
if domain == "" {
return nil, errors.New("Domain not passed as parameter or set in env USHER_DOMAIN")
}
// Set DBPath
dbpath := filepath.Join(root, domain+".yml")
// Set ConfigPath
configpath := filepath.Join(root, configfile)
return &DB{Root: root, Domain: domain, DBPath: dbpath, ConfigPath: configpath}, nil
}
// Init checks and creates the following, if they don't exist:
// - an usher root directory
// - an usher database for the db.Domain
// - an entry in the user config file for db.Domain
func (db *DB) Init() (dbCreated bool, err error) {
dbCreated = false
// Ensure root exists
err = os.MkdirAll(db.Root, 0755)
if err != nil {
return dbCreated, err
}
// Ensure database exists
_, err = os.Stat(db.DBPath)
if err == nil {
return dbCreated, nil // exists
}
if err != nil && !os.IsNotExist(err) {
return dbCreated, err // unexpected error
}
// Database does not exist - create
fh, err := os.Create(db.DBPath)
fh.Close()
if err != nil {
return dbCreated, err
}
dbCreated = true
// Ensure configfile exists
_, err = os.Stat(db.ConfigPath)
if err == nil {
_, err := db.readConfig()
if err != nil {
if err != ErrNotFound {
return dbCreated, err
}
}
err = db.appendConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
} else {
// Create a placeholder config file for domain
err = db.writeConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
}
return dbCreated, nil
}
// List returns the set of database entries whose code matches glob
func (db *DB) List(glob string) ([]Entry, error) {
// FIXME: first-pass - ignore glob
mappings, err := db.readDB()
if err != nil {
return nil, err
}
// Extract codes and sort
codes := make([]string, len(mappings))
i := 0
for code := range mappings {
codes[i] = code
i++
}
sort.Strings(codes)
// Compile entries
var entries = make([]Entry, len(mappings))
i = 0
for _, code := range codes {
entries[i] = Entry{Code: code, Url: mappings[code]}
i++
}
return entries, nil
}
// Add a mapping for url and code to the database.
// If code is missing, a random code will be generated and returned.
func (db *DB) Add(url, code string) (string, error) {
mappings, err := db.readDB()
if err != nil {
return "", err
}
if code == "" {
code = randomCode(mappings)
} else {
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// Check whether code is already used
dburl, exists := mappings[code]
if exists {
if dburl == url {
// Trying to re-add the same url is not an error, just a noop
return code, nil
}
return code, ErrCodeExists
}
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return code, err
}
return code, nil
}
// Update an existing mapping in the database, changing the URL.
func (db *DB) Update(url, code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// If code is missing, abort
dburl, exists := mappings[code]
if !exists {
return ErrNotFound
}
// Trying to update to the same url is not an error, just a noop
if dburl == url {
return nil
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Remove the mapping with code from the database
// Returns ErrNotFound if code does not exist in the database
func (db *DB) Remove(code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
_, exists := mappings[code]
if !exists {
return ErrNotFound
}
delete(mappings, code)
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Push syncs all current mappings with the backend configured for db.Domain
// in db.ConfigPath
func (db *DB) Push() error {
config, err := db.readConfig()
if err != nil {
return err
}
if config.Type == "" {
return fmt.Errorf("no 'type' field found for %q in config %q\n",
db.Domain, db.ConfigPath)
}
switch config.Type {
case "s3":
err = db.pushS3(config)
if err != nil {
return err
}
case "render":
err = db.pushRender()
if err != nil {
return err
}
case "unconfigured":
return ErrPushTypeUnconfigured
default:
return fmt.Errorf("invalid config backend type %q found for %q: %w",
config.Type, db.Domain, ErrPushTypeBad)
}
return nil
}
// readDB is a utility function to read all mappings from db.DBPath
// and return as a go map
func (db *DB) readDB() (map[string]string, error) {
data, err := ioutil.ReadFile(db.DBPath)
if err != nil {
return nil, err
}
var mappings map[string]string
err = yaml.Unmarshal(data, &mappings)
if err != nil {
return nil, err
}
if len(mappings) == 0 {
mappings = make(map[string]string)
}
return mappings, nil
}
// writeDB is a utility function to write mappings (as yaml) to db.DBPath
func (db *DB) writeDB(mappings map[string]string) error {
var data []byte
var err error
if len(mappings) > 0 {
data, err = yaml.Marshal(mappings)
if err != nil {
return err
}
}
tmpfile := db.DBPath + ".tmp"
err = ioutil.WriteFile(tmpfile, data, 0644)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.DBPath)
if err != nil {
return err
}
return nil
}
// readConfig is a utility function to read the config entry for
// db.Domain from db.ConfigPath file
func (db *DB) readConfig() (*ConfigEntry, error) {
data, err := ioutil.ReadFile(db
|
{
cwd, err := os.Getwd()
if err == nil {
root = cwd
}
}
|
conditional_block
|
usher.go
|
Dir, "usher")
}
// Derive domain if not set - check for USHER_DOMAIN in environment
if domain == "" {
domain = os.Getenv("USHER_DOMAIN")
}
// Else infer the domain if only one database exists
if domain == "" {
matches, _ := filepath.Glob(filepath.Join(root, "*.*.yml"))
if len(matches) == 1 {
// Exactly one match - strip .yml suffix to get domain
re := regexp.MustCompile(`.yml$`)
domain = re.ReplaceAllLiteralString(filepath.Base(matches[0]), "")
}
}
// Else give up with an error
if domain == "" {
return nil, errors.New("Domain not passed as parameter or set in env USHER_DOMAIN")
}
// Set DBPath
dbpath := filepath.Join(root, domain+".yml")
// Set ConfigPath
configpath := filepath.Join(root, configfile)
return &DB{Root: root, Domain: domain, DBPath: dbpath, ConfigPath: configpath}, nil
}
// Init checks and creates the following, if they don't exist:
// - an usher root directory
// - an usher database for the db.Domain
// - an entry in the user config file for db.Domain
func (db *DB) Init() (dbCreated bool, err error) {
dbCreated = false
// Ensure root exists
err = os.MkdirAll(db.Root, 0755)
if err != nil {
return dbCreated, err
}
// Ensure database exists
_, err = os.Stat(db.DBPath)
if err == nil {
return dbCreated, nil // exists
}
if err != nil && !os.IsNotExist(err) {
return dbCreated, err // unexpected error
}
// Database does not exist - create
fh, err := os.Create(db.DBPath)
fh.Close()
if err != nil {
return dbCreated, err
}
dbCreated = true
// Ensure configfile exists
_, err = os.Stat(db.ConfigPath)
if err == nil {
_, err := db.readConfig()
if err != nil {
if err != ErrNotFound {
return dbCreated, err
}
}
err = db.appendConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
} else {
// Create a placeholder config file for domain
err = db.writeConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
}
return dbCreated, nil
}
// List returns the set of database entries whose code matches glob
func (db *DB) List(glob string) ([]Entry, error) {
// FIXME: first-pass - ignore glob
mappings, err := db.readDB()
if err != nil {
return nil, err
}
// Extract codes and sort
codes := make([]string, len(mappings))
i := 0
for code := range mappings {
codes[i] = code
i++
}
sort.Strings(codes)
// Compile entries
var entries = make([]Entry, len(mappings))
i = 0
for _, code := range codes {
entries[i] = Entry{Code: code, Url: mappings[code]}
i++
}
return entries, nil
}
// Add a mapping for url and code to the database.
// If code is missing, a random code will be generated and returned.
func (db *DB) Add(url, code string) (string, error) {
mappings, err := db.readDB()
if err != nil {
return "", err
}
if code == "" {
code = randomCode(mappings)
} else {
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// Check whether code is already used
dburl, exists := mappings[code]
if exists {
if dburl == url {
// Trying to re-add the same url is not an error, just a noop
return code, nil
}
return code, ErrCodeExists
}
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return code, err
}
return code, nil
}
// Update an existing mapping in the database, changing the URL.
func (db *DB) Update(url, code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// If code is missing, abort
dburl, exists := mappings[code]
if !exists {
return ErrNotFound
}
// Trying to update to the same url is not an error, just a noop
if dburl == url {
return nil
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Remove the mapping with code from the database
// Returns ErrNotFound if code does not exist in the database
func (db *DB) Remove(code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
_, exists := mappings[code]
if !exists {
return ErrNotFound
}
delete(mappings, code)
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Push syncs all current mappings with the backend configured for db.Domain
// in db.ConfigPath
func (db *DB) Push() error {
config, err := db.readConfig()
if err != nil {
return err
}
if config.Type == "" {
return fmt.Errorf("no 'type' field found for %q in config %q\n",
db.Domain, db.ConfigPath)
}
switch config.Type {
case "s3":
err = db.pushS3(config)
if err != nil {
return err
}
case "render":
err = db.pushRender()
if err != nil {
return err
}
case "unconfigured":
return ErrPushTypeUnconfigured
default:
return fmt.Errorf("invalid config backend type %q found for %q: %w",
config.Type, db.Domain, ErrPushTypeBad)
}
return nil
}
// readDB is a utility function to read all mappings from db.DBPath
// and return as a go map
func (db *DB) readDB() (map[string]string, error) {
data, err := ioutil.ReadFile(db.DBPath)
if err != nil {
return nil, err
}
var mappings map[string]string
err = yaml.Unmarshal(data, &mappings)
if err != nil {
return nil, err
}
if len(mappings) == 0 {
mappings = make(map[string]string)
}
return mappings, nil
}
// writeDB is a utility function to write mappings (as yaml) to db.DBPath
func (db *DB) writeDB(mappings map[string]string) error {
var data []byte
var err error
if len(mappings) > 0 {
data, err = yaml.Marshal(mappings)
if err != nil {
return err
}
}
tmpfile := db.DBPath + ".tmp"
err = ioutil.WriteFile(tmpfile, data, 0644)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.DBPath)
if err != nil {
return err
}
return nil
}
// readConfig is a utility function to read the config entry for
// db.Domain from db.ConfigPath file
func (db *DB) readConfig() (*ConfigEntry, error) {
data, err := ioutil.ReadFile(db.ConfigPath)
if err != nil {
return nil, err
}
var entries map[string]ConfigEntry
err = yaml.Unmarshal(data, &entries)
if err != nil {
return nil, err
}
entry, exists := entries[db.Domain]
if !exists {
return nil, ErrNotFound
}
return &entry, nil
}
// writeConfigString is a utility function to write data to db.ConfigPath
func (db *DB) writeConfigString(data string) error {
tmpfile := db.ConfigPath + ".tmp"
err := ioutil.WriteFile(tmpfile, []byte(data), 0600)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.ConfigPath)
if err != nil {
return err
}
return nil
}
// appendConfigString is a utility function to write data to db.ConfigPath
func (db *DB) appendConfigString(data string) error
|
{
config, err := ioutil.ReadFile(db.ConfigPath)
if err != nil {
return err
}
config = append(config, []byte(data)...)
tmpfile := db.ConfigPath + ".tmp"
err = ioutil.WriteFile(tmpfile, config, 0600)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.ConfigPath)
if err != nil {
return err
}
return nil
}
|
identifier_body
|
|
usher.go
|
-qualified domain whose mappings we want
DBPath string // full path to database for Domain
ConfigPath string // full path to usher config file
}
type Entry struct {
Code string
Url string
}
type ConfigEntry struct {
Type string `yaml:"type"`
AWSKey string `yaml:"aws_key,omitempty"`
AWSSecret string `yaml:"aws_secret,omitempty"`
AWSRegion string `yaml:"aws_region,omitempty"`
}
// NewDB creates a DB struct with members derived from parameters,
// the environment, or defaults (in that order). It does no checking
// that the values produced are sane or exist on the filesystem.
func NewDB(domain string) (*DB, error) {
// Get root
root := os.Getenv("USHER_ROOT")
if root == "" {
// If USHER_ROOT is unset, check if there is an usher.yml in the cwd
stat, err := os.Stat("usher.yml")
if err == nil && !stat.IsDir() {
cwd, err := os.Getwd()
if err == nil {
root = cwd
}
}
}
if root == "" {
// If root is still unset, default to "os.UserConfigDir()/usher"
configDir, err := os.UserConfigDir()
if err != nil {
return nil, err
}
root = filepath.Join(configDir, "usher")
}
// Derive domain if not set - check for USHER_DOMAIN in environment
if domain == "" {
domain = os.Getenv("USHER_DOMAIN")
}
// Else infer the domain if only one database exists
if domain == "" {
matches, _ := filepath.Glob(filepath.Join(root, "*.*.yml"))
if len(matches) == 1 {
// Exactly one match - strip .yml suffix to get domain
re := regexp.MustCompile(`.yml$`)
domain = re.ReplaceAllLiteralString(filepath.Base(matches[0]), "")
}
}
// Else give up with an error
if domain == "" {
return nil, errors.New("Domain not passed as parameter or set in env USHER_DOMAIN")
}
// Set DBPath
dbpath := filepath.Join(root, domain+".yml")
// Set ConfigPath
configpath := filepath.Join(root, configfile)
return &DB{Root: root, Domain: domain, DBPath: dbpath, ConfigPath: configpath}, nil
}
// Init checks and creates the following, if they don't exist:
// - an usher root directory
// - an usher database for the db.Domain
// - an entry in the user config file for db.Domain
func (db *DB) Init() (dbCreated bool, err error) {
dbCreated = false
// Ensure root exists
err = os.MkdirAll(db.Root, 0755)
if err != nil {
return dbCreated, err
}
// Ensure database exists
_, err = os.Stat(db.DBPath)
if err == nil {
return dbCreated, nil // exists
}
if err != nil && !os.IsNotExist(err) {
return dbCreated, err // unexpected error
}
// Database does not exist - create
fh, err := os.Create(db.DBPath)
fh.Close()
if err != nil {
return dbCreated, err
}
dbCreated = true
// Ensure configfile exists
_, err = os.Stat(db.ConfigPath)
if err == nil {
_, err := db.readConfig()
if err != nil {
if err != ErrNotFound {
return dbCreated, err
}
}
err = db.appendConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
} else {
// Create a placeholder config file for domain
err = db.writeConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
}
return dbCreated, nil
}
// List returns the set of database entries whose code matches glob
func (db *DB) List(glob string) ([]Entry, error) {
// FIXME: first-pass - ignore glob
mappings, err := db.readDB()
if err != nil {
return nil, err
}
// Extract codes and sort
codes := make([]string, len(mappings))
i := 0
for code := range mappings {
|
// Compile entries
var entries = make([]Entry, len(mappings))
i = 0
for _, code := range codes {
entries[i] = Entry{Code: code, Url: mappings[code]}
i++
}
return entries, nil
}
// Add a mapping for url and code to the database.
// If code is missing, a random code will be generated and returned.
func (db *DB) Add(url, code string) (string, error) {
mappings, err := db.readDB()
if err != nil {
return "", err
}
if code == "" {
code = randomCode(mappings)
} else {
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// Check whether code is already used
dburl, exists := mappings[code]
if exists {
if dburl == url {
// Trying to re-add the same url is not an error, just a noop
return code, nil
}
return code, ErrCodeExists
}
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return code, err
}
return code, nil
}
// Update an existing mapping in the database, changing the URL.
func (db *DB) Update(url, code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// If code is missing, abort
dburl, exists := mappings[code]
if !exists {
return ErrNotFound
}
// Trying to update to the same url is not an error, just a noop
if dburl == url {
return nil
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Remove the mapping with code from the database
// Returns ErrNotFound if code does not exist in the database
func (db *DB) Remove(code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
_, exists := mappings[code]
if !exists {
return ErrNotFound
}
delete(mappings, code)
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Push syncs all current mappings with the backend configured for db.Domain
// in db.ConfigPath
func (db *DB) Push() error {
config, err := db.readConfig()
if err != nil {
return err
}
if config.Type == "" {
return fmt.Errorf("no 'type' field found for %q in config %q\n",
db.Domain, db.ConfigPath)
}
switch config.Type {
case "s3":
err = db.pushS3(config)
if err != nil {
return err
}
case "render":
err = db.pushRender()
if err != nil {
return err
}
case "unconfigured":
return ErrPushTypeUnconfigured
default:
return fmt.Errorf("invalid config backend type %q found for %q: %w",
config.Type, db.Domain, ErrPushTypeBad)
}
return nil
}
// readDB is a utility function to read all mappings from db.DBPath
// and return as a go map
func (db *DB) readDB() (map[string]string, error) {
data, err := ioutil.ReadFile(db.DBPath)
if err != nil {
return nil, err
}
var mappings map[string]string
err = yaml.Unmarshal(data, &mappings)
if err != nil {
return nil, err
}
if len(mappings) == 0 {
mappings = make(map[string]string)
}
return mappings, nil
}
// writeDB is a utility function to write mappings (as yaml) to db.DBPath
func (db *DB) writeDB(mappings map[string]string) error {
var data []byte
var err error
if len(mappings) > 0 {
data, err = yaml.Marshal(mappings)
if err != nil {
return err
}
}
tmpfile := db.DBPath + ".tmp"
err = ioutil.WriteFile(tmpfile, data, 0644)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.DBPath)
if err != nil {
return err
}
return nil
}
// readConfig is a utility function to read the config entry for
// db.Domain from db.ConfigPath file
func (db *DB) readConfig() (*ConfigEntry, error) {
data, err := ioutil.ReadFile(db.Config
|
codes[i] = code
i++
}
sort.Strings(codes)
|
random_line_split
|
usher.go
|
yml"))
if len(matches) == 1 {
// Exactly one match - strip .yml suffix to get domain
re := regexp.MustCompile(`.yml$`)
domain = re.ReplaceAllLiteralString(filepath.Base(matches[0]), "")
}
}
// Else give up with an error
if domain == "" {
return nil, errors.New("Domain not passed as parameter or set in env USHER_DOMAIN")
}
// Set DBPath
dbpath := filepath.Join(root, domain+".yml")
// Set ConfigPath
configpath := filepath.Join(root, configfile)
return &DB{Root: root, Domain: domain, DBPath: dbpath, ConfigPath: configpath}, nil
}
// Init checks and creates the following, if they don't exist:
// - an usher root directory
// - an usher database for the db.Domain
// - an entry in the user config file for db.Domain
func (db *DB) Init() (dbCreated bool, err error) {
dbCreated = false
// Ensure root exists
err = os.MkdirAll(db.Root, 0755)
if err != nil {
return dbCreated, err
}
// Ensure database exists
_, err = os.Stat(db.DBPath)
if err == nil {
return dbCreated, nil // exists
}
if err != nil && !os.IsNotExist(err) {
return dbCreated, err // unexpected error
}
// Database does not exist - create
fh, err := os.Create(db.DBPath)
fh.Close()
if err != nil {
return dbCreated, err
}
dbCreated = true
// Ensure configfile exists
_, err = os.Stat(db.ConfigPath)
if err == nil {
_, err := db.readConfig()
if err != nil {
if err != ErrNotFound {
return dbCreated, err
}
}
err = db.appendConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
} else {
// Create a placeholder config file for domain
err = db.writeConfigString(db.configPlaceholder())
if err != nil {
return dbCreated, err
}
}
return dbCreated, nil
}
// List returns the set of database entries whose code matches glob
func (db *DB) List(glob string) ([]Entry, error) {
// FIXME: first-pass - ignore glob
mappings, err := db.readDB()
if err != nil {
return nil, err
}
// Extract codes and sort
codes := make([]string, len(mappings))
i := 0
for code := range mappings {
codes[i] = code
i++
}
sort.Strings(codes)
// Compile entries
var entries = make([]Entry, len(mappings))
i = 0
for _, code := range codes {
entries[i] = Entry{Code: code, Url: mappings[code]}
i++
}
return entries, nil
}
// Add a mapping for url and code to the database.
// If code is missing, a random code will be generated and returned.
func (db *DB) Add(url, code string) (string, error) {
mappings, err := db.readDB()
if err != nil {
return "", err
}
if code == "" {
code = randomCode(mappings)
} else {
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// Check whether code is already used
dburl, exists := mappings[code]
if exists {
if dburl == url {
// Trying to re-add the same url is not an error, just a noop
return code, nil
}
return code, ErrCodeExists
}
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return code, err
}
return code, nil
}
// Update an existing mapping in the database, changing the URL.
func (db *DB) Update(url, code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
// Check for parameter inversion
reUrl := regexp.MustCompile(`^https?://`)
if !reUrl.MatchString(url) && reUrl.MatchString(code) {
url, code = code, url
}
// If code is missing, abort
dburl, exists := mappings[code]
if !exists {
return ErrNotFound
}
// Trying to update to the same url is not an error, just a noop
if dburl == url {
return nil
}
mappings[code] = url
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Remove the mapping with code from the database
// Returns ErrNotFound if code does not exist in the database
func (db *DB) Remove(code string) error {
mappings, err := db.readDB()
if err != nil {
return err
}
_, exists := mappings[code]
if !exists {
return ErrNotFound
}
delete(mappings, code)
err = db.writeDB(mappings)
if err != nil {
return err
}
return nil
}
// Push syncs all current mappings with the backend configured for db.Domain
// in db.ConfigPath
func (db *DB) Push() error {
config, err := db.readConfig()
if err != nil {
return err
}
if config.Type == "" {
return fmt.Errorf("no 'type' field found for %q in config %q\n",
db.Domain, db.ConfigPath)
}
switch config.Type {
case "s3":
err = db.pushS3(config)
if err != nil {
return err
}
case "render":
err = db.pushRender()
if err != nil {
return err
}
case "unconfigured":
return ErrPushTypeUnconfigured
default:
return fmt.Errorf("invalid config backend type %q found for %q: %w",
config.Type, db.Domain, ErrPushTypeBad)
}
return nil
}
// readDB is a utility function to read all mappings from db.DBPath
// and return as a go map
func (db *DB) readDB() (map[string]string, error) {
data, err := ioutil.ReadFile(db.DBPath)
if err != nil {
return nil, err
}
var mappings map[string]string
err = yaml.Unmarshal(data, &mappings)
if err != nil {
return nil, err
}
if len(mappings) == 0 {
mappings = make(map[string]string)
}
return mappings, nil
}
// writeDB is a utility function to write mappings (as yaml) to db.DBPath
func (db *DB) writeDB(mappings map[string]string) error {
var data []byte
var err error
if len(mappings) > 0 {
data, err = yaml.Marshal(mappings)
if err != nil {
return err
}
}
tmpfile := db.DBPath + ".tmp"
err = ioutil.WriteFile(tmpfile, data, 0644)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.DBPath)
if err != nil {
return err
}
return nil
}
// readConfig is a utility function to read the config entry for
// db.Domain from db.ConfigPath file
func (db *DB) readConfig() (*ConfigEntry, error) {
data, err := ioutil.ReadFile(db.ConfigPath)
if err != nil {
return nil, err
}
var entries map[string]ConfigEntry
err = yaml.Unmarshal(data, &entries)
if err != nil {
return nil, err
}
entry, exists := entries[db.Domain]
if !exists {
return nil, ErrNotFound
}
return &entry, nil
}
// writeConfigString is a utility function to write data to db.ConfigPath
func (db *DB) writeConfigString(data string) error {
tmpfile := db.ConfigPath + ".tmp"
err := ioutil.WriteFile(tmpfile, []byte(data), 0600)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.ConfigPath)
if err != nil {
return err
}
return nil
}
// appendConfigString is a utility function to write data to db.ConfigPath
func (db *DB) appendConfigString(data string) error {
config, err := ioutil.ReadFile(db.ConfigPath)
if err != nil {
return err
}
config = append(config, []byte(data)...)
tmpfile := db.ConfigPath + ".tmp"
err = ioutil.WriteFile(tmpfile, config, 0600)
if err != nil {
return err
}
err = os.Rename(tmpfile, db.ConfigPath)
if err != nil {
return err
}
return nil
}
// randomCode is a utility function to generate a random code
// and check that it doesn't exist in mappings.
// Random codes use the following pattern: 1 digit, then 4-7
// lowercase ascii characters. This usually allows them to be
// relatively easily distinguished from explicit codes, while
// still being easy to communicate orally.
func
|
randomCode
|
identifier_name
|
|
forall.rs
|
self.state;
let m = &mut self.max;
let mut i = 0;
loop {
if i == v.len()-1 {
|
return None;
}
if v[i] > 0 {
v[i+1] += 1;
v[i] -= 1;
if v[i+1] <= m[i+1] {
break;
}
}
i += 1;
}
let mut res = v[0..=i].iter().sum();
let mut j = 0;
while res > 0 {
let cur = std::cmp::min(m[j],res);
v[j] = cur;
res -= cur;
j += 1;
}
for k in j..=i {
v[k] = 0;
}
return Some(&self.state);
}
}
}
struct Matches {
state : Vec<Comb>,
first : bool,
v1 : Vec<usize>
}
impl Matches {
fn new(v1 : Vec<usize>, mut v2 : Vec<usize>) -> Self {
let mut s = vec![];
for &x in &v1 {
let mut c = Comb::new(x,v2.clone());
c.next();
for i in 0..v2.len() {
v2[i] -= c.state[i];
}
s.push(c);
}
Self {
v1, state : s, first : true
}
}
fn next(&mut self) -> Option<&Vec<Comb>> {
if self.first {
self.first = false;
Some(&self.state)
}else {
for i in (0..self.state.len()).rev() {
if self.state[i].next() != None {
for j in i+1..self.state.len() {
let split = self.state.split_at_mut(j);
let p = &split.0[j-1];
let p2 = &mut split.1[0];
let pmax = &p.max;
let ps = &p.state;
let n = self.v1[j];
p2.transform(n,pmax.iter().zip(ps.iter()).map(|(m,x)|m-x));
//let v : Vec<_> = pmax.iter().zip(ps.iter()).map(|(m,x)|m-x).collect();
//self.state[j] = Comb::new(n,v);
p2.next();
}
return Some(&self.state);
}
}
None
}
}
}
fn count_map<BigNum>(v : &[BigNum]) -> HashMap<BigNum,usize> where BigNum : crate::bignum::BigNum{
let mut h = HashMap::new();
for n in v {
*h.entry(n.clone()).or_default() += 1;
}
h
}
#[inline(never)]
fn intersections<BigNum>(uni : BigNum, c1 : &[(BigNum,usize)], c2 : &[(BigNum,usize)], delta : usize , bits : usize) -> Vec<Line<BigNum>> where BigNum : crate::bignum::BigNum {
let v1 : Vec<_> = c1.iter().map(|(_,c)|*c).collect();
let v2 : Vec<_> = c2.iter().map(|(_,c)|*c).collect();
let mut m = Matches::new(v1,v2);
let mut r = vec![];
let mut oldbad : Option<(usize,usize,usize,usize)> = None;
'outer: while let Some(x) = m.next() {
if let Some((i1,i2,j1,j2)) = oldbad {
if x[i1].state[j1] != 0 && x[i2].state[j2] != 0 {
continue 'outer;
}
}
for i1 in 0..c1.len() {
for j1 in 0..c2.len() {
if x[i1].state[j1] != 0 {
for i2 in i1+1..c1.len() {
for j2 in 0..c2.len() {
if x[i2].state[j2] != 0 {
let u1 = c1[i1].0.clone() & c2[j1].0.clone();
let u2 = c1[i2].0.clone() & c2[j2].0.clone();
let u3 = c1[i1].0.clone() & c2[j2].0.clone();
let u4 = c1[i2].0.clone() & c2[j1].0.clone();
if (u4.is_superset(&u1) && u3.is_superset(&u2) && (u1 != u4 || u2 != u3)) || (u3.is_superset(&u1) && u4.is_superset(&u2) && (u1 != u3 || u2 != u4)) {
oldbad = Some((i1,i2,j1,j2));
continue 'outer;
}
}
}
}
}
}
}
let mut groups = Vec::with_capacity(delta);
groups.push(uni.clone());
for (i,(ga,_)) in c1.iter().enumerate() {
for (j,(gb,_)) in c2.iter().enumerate() {
for _ in 0..x[i].state[j] {
groups.push(ga.clone() & gb.clone());
}
}
}
if !groups.contains(&BigNum::zero()) {
r.push(Line::from_groups(delta, bits, groups.into_iter()).sorted());
}
}
r
}
#[inline(never)]
fn perm_includes<BigNum>(line : &Line<BigNum>, other : &Line<BigNum>) -> bool where BigNum : crate::bignum::BigNum {
let g1 : Vec<_> = line.groups().collect();
let g2 : Vec<_> = other.groups().collect();
let d = g1.len();
let mut g = contest_algorithms::graph::flow::FlowGraph::new(2*d+2,d*d);
for i in 1..=d {
g.add_edge(0, i, 1, 0, 0);
}
for i in d+1..=2*d {
g.add_edge(i, 2*d+1, 1, 0, 0);
}
for i in 0..d {
for j in 0..d {
if g1[i].is_superset(&g2[j]) {
g.add_edge(1+i, 1+d+j, 1, 0, 0);
}
}
}
g.dinic(0, 2*d+1).0 == d as i64
}
#[inline(never)]
fn add_reduce_maximal<BigNum>(lines : &mut Vec<Line<BigNum>>, newline : Line<BigNum>) where BigNum : crate::bignum::BigNum {
let l1 = lines.len();
lines.retain(|oldline| !perm_includes(&newline, oldline));
let l2 = lines.len();
if l1 != l2 || lines.iter().all(|oldline|!perm_includes(oldline,&newline)) {
lines.push(newline);
}
}
#[inline(never)]
fn find_good_unions<BigNum>(u1 : &[BigNum], u2 : &[BigNum]) -> HashMap<BigNum,Vec<(BigNum,BigNum)>> where BigNum : crate::bignum::BigNum {
let mut unions = HashMap::new();
for x in u1.iter() {
for y in u2.iter() {
if x.is_superset(y) || y.is_superset(x) {
continue;
}
let uni = x.clone() | y.clone();
let unis : &mut Vec<(BigNum,BigNum)> = unions.entry(uni).or_insert(vec![]);
let len = unis.len();
unis.retain(|(xc,yc)| !(xc.is_superset(x) && yc.is_superset(y)) );
if unis.len() != len || unis.iter().all(|(xc,yc)| !(x.is_superset(xc) && y.is_superset(yc)) ) {
unis.push((x.clone(),y.clone()));
}
}
}
unions
}
pub fn forall<BigNum>(nc : &Constraint<BigNum>, problem : &Problem<BigNum>) -> Constraint<BigNum> where BigNum : crate::bignum::BigNum {
let mut nc = nc.clone();
let maplt = problem.map_label_text();
let set_to_string = |s:&BigNum|{
let r = s.one_bits().map(|elem|&maplt[&elem]).join("");
if r == "" {
String::from("_")
}else{
r
}
};
let make_right_closed = |g : BigNum|{g.clone()|problem.successors(g.one_bits().next().unwrap(),false)};
for line in &mut nc.lines {
*line = line.edited(|g|{make_right_closed(g)}).sorted();
}
let mut seen : HashSet<_> = nc.lines.iter().cloned().collect();
let lines = std::mem::replace(&mut nc.lines, vec![]);
for line in lines {
seen.insert(line.clone());
add_reduce
|
random_line_split
|
|
forall.rs
|
self.state;
let m = &mut self.max;
let mut i = 0;
loop {
if i == v.len()-1 {
return None;
}
if v[i] > 0 {
v[i+1] += 1;
v[i] -= 1;
if v[i+1] <= m[i+1] {
break;
}
}
i += 1;
}
let mut res = v[0..=i].iter().sum();
let mut j = 0;
while res > 0 {
let cur = std::cmp::min(m[j],res);
v[j] = cur;
res -= cur;
j += 1;
}
for k in j..=i {
v[k] = 0;
}
return Some(&self.state);
}
}
}
struct Matches {
state : Vec<Comb>,
first : bool,
v1 : Vec<usize>
}
impl Matches {
fn new(v1 : Vec<usize>, mut v2 : Vec<usize>) -> Self {
let mut s = vec![];
for &x in &v1 {
let mut c = Comb::new(x,v2.clone());
c.next();
for i in 0..v2.len() {
v2[i] -= c.state[i];
}
s.push(c);
}
Self {
v1, state : s, first : true
}
}
fn next(&mut self) -> Option<&Vec<Comb>> {
if self.first {
self.first = false;
Some(&self.state)
}else {
for i in (0..self.state.len()).rev() {
if self.state[i].next() != None {
for j in i+1..self.state.len() {
let split = self.state.split_at_mut(j);
let p = &split.0[j-1];
let p2 = &mut split.1[0];
let pmax = &p.max;
let ps = &p.state;
let n = self.v1[j];
p2.transform(n,pmax.iter().zip(ps.iter()).map(|(m,x)|m-x));
//let v : Vec<_> = pmax.iter().zip(ps.iter()).map(|(m,x)|m-x).collect();
//self.state[j] = Comb::new(n,v);
p2.next();
}
return Some(&self.state);
}
}
None
}
}
}
fn count_map<BigNum>(v : &[BigNum]) -> HashMap<BigNum,usize> where BigNum : crate::bignum::BigNum{
let mut h = HashMap::new();
for n in v {
*h.entry(n.clone()).or_default() += 1;
}
h
}
#[inline(never)]
fn intersections<BigNum>(uni : BigNum, c1 : &[(BigNum,usize)], c2 : &[(BigNum,usize)], delta : usize , bits : usize) -> Vec<Line<BigNum>> where BigNum : crate::bignum::BigNum {
let v1 : Vec<_> = c1.iter().map(|(_,c)|*c).collect();
let v2 : Vec<_> = c2.iter().map(|(_,c)|*c).collect();
let mut m = Matches::new(v1,v2);
let mut r = vec![];
let mut oldbad : Option<(usize,usize,usize,usize)> = None;
'outer: while let Some(x) = m.next() {
if let Some((i1,i2,j1,j2)) = oldbad {
if x[i1].state[j1] != 0 && x[i2].state[j2] != 0 {
continue 'outer;
}
}
for i1 in 0..c1.len() {
for j1 in 0..c2.len() {
if x[i1].state[j1] != 0 {
for i2 in i1+1..c1.len() {
for j2 in 0..c2.len() {
if x[i2].state[j2] != 0
|
}
}
}
}
}
let mut groups = Vec::with_capacity(delta);
groups.push(uni.clone());
for (i,(ga,_)) in c1.iter().enumerate() {
for (j,(gb,_)) in c2.iter().enumerate() {
for _ in 0..x[i].state[j] {
groups.push(ga.clone() & gb.clone());
}
}
}
if !groups.contains(&BigNum::zero()) {
r.push(Line::from_groups(delta, bits, groups.into_iter()).sorted());
}
}
r
}
#[inline(never)]
fn perm_includes<BigNum>(line : &Line<BigNum>, other : &Line<BigNum>) -> bool where BigNum : crate::bignum::BigNum {
let g1 : Vec<_> = line.groups().collect();
let g2 : Vec<_> = other.groups().collect();
let d = g1.len();
let mut g = contest_algorithms::graph::flow::FlowGraph::new(2*d+2,d*d);
for i in 1..=d {
g.add_edge(0, i, 1, 0, 0);
}
for i in d+1..=2*d {
g.add_edge(i, 2*d+1, 1, 0, 0);
}
for i in 0..d {
for j in 0..d {
if g1[i].is_superset(&g2[j]) {
g.add_edge(1+i, 1+d+j, 1, 0, 0);
}
}
}
g.dinic(0, 2*d+1).0 == d as i64
}
#[inline(never)]
fn add_reduce_maximal<BigNum>(lines : &mut Vec<Line<BigNum>>, newline : Line<BigNum>) where BigNum : crate::bignum::BigNum {
let l1 = lines.len();
lines.retain(|oldline| !perm_includes(&newline, oldline));
let l2 = lines.len();
if l1 != l2 || lines.iter().all(|oldline|!perm_includes(oldline,&newline)) {
lines.push(newline);
}
}
#[inline(never)]
fn find_good_unions<BigNum>(u1 : &[BigNum], u2 : &[BigNum]) -> HashMap<BigNum,Vec<(BigNum,BigNum)>> where BigNum : crate::bignum::BigNum {
let mut unions = HashMap::new();
for x in u1.iter() {
for y in u2.iter() {
if x.is_superset(y) || y.is_superset(x) {
continue;
}
let uni = x.clone() | y.clone();
let unis : &mut Vec<(BigNum,BigNum)> = unions.entry(uni).or_insert(vec![]);
let len = unis.len();
unis.retain(|(xc,yc)| !(xc.is_superset(x) && yc.is_superset(y)) );
if unis.len() != len || unis.iter().all(|(xc,yc)| !(x.is_superset(xc) && y.is_superset(yc)) ) {
unis.push((x.clone(),y.clone()));
}
}
}
unions
}
pub fn forall<BigNum>(nc : &Constraint<BigNum>, problem : &Problem<BigNum>) -> Constraint<BigNum> where BigNum : crate::bignum::BigNum {
let mut nc = nc.clone();
let maplt = problem.map_label_text();
let set_to_string = |s:&BigNum|{
let r = s.one_bits().map(|elem|&maplt[&elem]).join("");
if r == "" {
String::from("_")
}else{
r
}
};
let make_right_closed = |g : BigNum|{g.clone()|problem.successors(g.one_bits().next().unwrap(),false)};
for line in &mut nc.lines {
*line = line.edited(|g|{make_right_closed(g)}).sorted();
}
let mut seen : HashSet<_> = nc.lines.iter().cloned().collect();
let lines = std::mem::replace(&mut nc.lines, vec![]);
for line in lines {
seen.insert(line.clone());
|
{
let u1 = c1[i1].0.clone() & c2[j1].0.clone();
let u2 = c1[i2].0.clone() & c2[j2].0.clone();
let u3 = c1[i1].0.clone() & c2[j2].0.clone();
let u4 = c1[i2].0.clone() & c2[j1].0.clone();
if (u4.is_superset(&u1) && u3.is_superset(&u2) && (u1 != u4 || u2 != u3)) || (u3.is_superset(&u1) && u4.is_superset(&u2) && (u1 != u3 || u2 != u4)) {
oldbad = Some((i1,i2,j1,j2));
continue 'outer;
}
}
|
conditional_block
|
forall.rs
|
fn transform(&mut self, n : usize, max : impl Iterator<Item=usize>) {
let mut i = 0;
for x in max {
self.max[i] = x;
i += 1;
}
assert!(i == self.max.len());
let mut res = n;
let mut i = 0;
while res > 0 {
let cur = std::cmp::min(self.max[i],res);
self.state[i] = cur;
res -= cur;
i += 1;
}
for j in i..self.state.len() {
self.state[j] = 0;
}
self.first = true;
}
fn next(&mut self) -> Option<&Vec<usize>> {
if self.first {
self.first = false;
Some(&self.state)
}else {
let v = &mut self.state;
let m = &mut self.max;
let mut i = 0;
loop {
if i == v.len()-1 {
return None;
}
if v[i] > 0 {
v[i+1] += 1;
v[i] -= 1;
if v[i+1] <= m[i+1] {
break;
}
}
i += 1;
}
let mut res = v[0..=i].iter().sum();
let mut j = 0;
while res > 0 {
let cur = std::cmp::min(m[j],res);
v[j] = cur;
res -= cur;
j += 1;
}
for k in j..=i {
v[k] = 0;
}
return Some(&self.state);
}
}
}
struct Matches {
state : Vec<Comb>,
first : bool,
v1 : Vec<usize>
}
impl Matches {
fn new(v1 : Vec<usize>, mut v2 : Vec<usize>) -> Self {
let mut s = vec![];
for &x in &v1 {
let mut c = Comb::new(x,v2.clone());
c.next();
for i in 0..v2.len() {
v2[i] -= c.state[i];
}
s.push(c);
}
Self {
v1, state : s, first : true
}
}
fn next(&mut self) -> Option<&Vec<Comb>> {
if self.first {
self.first = false;
Some(&self.state)
}else {
for i in (0..self.state.len()).rev() {
if self.state[i].next() != None {
for j in i+1..self.state.len() {
let split = self.state.split_at_mut(j);
let p = &split.0[j-1];
let p2 = &mut split.1[0];
let pmax = &p.max;
let ps = &p.state;
let n = self.v1[j];
p2.transform(n,pmax.iter().zip(ps.iter()).map(|(m,x)|m-x));
//let v : Vec<_> = pmax.iter().zip(ps.iter()).map(|(m,x)|m-x).collect();
//self.state[j] = Comb::new(n,v);
p2.next();
}
return Some(&self.state);
}
}
None
}
}
}
fn count_map<BigNum>(v : &[BigNum]) -> HashMap<BigNum,usize> where BigNum : crate::bignum::BigNum{
let mut h = HashMap::new();
for n in v {
*h.entry(n.clone()).or_default() += 1;
}
h
}
#[inline(never)]
fn intersections<BigNum>(uni : BigNum, c1 : &[(BigNum,usize)], c2 : &[(BigNum,usize)], delta : usize , bits : usize) -> Vec<Line<BigNum>> where BigNum : crate::bignum::BigNum {
let v1 : Vec<_> = c1.iter().map(|(_,c)|*c).collect();
let v2 : Vec<_> = c2.iter().map(|(_,c)|*c).collect();
let mut m = Matches::new(v1,v2);
let mut r = vec![];
let mut oldbad : Option<(usize,usize,usize,usize)> = None;
'outer: while let Some(x) = m.next() {
if let Some((i1,i2,j1,j2)) = oldbad {
if x[i1].state[j1] != 0 && x[i2].state[j2] != 0 {
continue 'outer;
}
}
for i1 in 0..c1.len() {
for j1 in 0..c2.len() {
if x[i1].state[j1] != 0 {
for i2 in i1+1..c1.len() {
for j2 in 0..c2.len() {
if x[i2].state[j2] != 0 {
let u1 = c1[i1].0.clone() & c2[j1].0.clone();
let u2 = c1[i2].0.clone() & c2[j2].0.clone();
let u3 = c1[i1].0.clone() & c2[j2].0.clone();
let u4 = c1[i2].0.clone() & c2[j1].0.clone();
if (u4.is_superset(&u1) && u3.is_superset(&u2) && (u1 != u4 || u2 != u3)) || (u3.is_superset(&u1) && u4.is_superset(&u2) && (u1 != u3 || u2 != u4)) {
oldbad = Some((i1,i2,j1,j2));
continue 'outer;
}
}
}
}
}
}
}
let mut groups = Vec::with_capacity(delta);
groups.push(uni.clone());
for (i,(ga,_)) in c1.iter().enumerate() {
for (j,(gb,_)) in c2.iter().enumerate() {
for _ in 0..x[i].state[j] {
groups.push(ga.clone() & gb.clone());
}
}
}
if !groups.contains(&BigNum::zero()) {
r.push(Line::from_groups(delta, bits, groups.into_iter()).sorted());
}
}
r
}
#[inline(never)]
fn perm_includes<BigNum>(line : &Line<BigNum>, other : &Line<BigNum>) -> bool where BigNum : crate::bignum::BigNum {
let g1 : Vec<_> = line.groups().collect();
let g2 : Vec<_> = other.groups().collect();
let d = g1.len();
let mut g = contest_algorithms::graph::flow::FlowGraph::new(2*d+2,d*d);
for i in 1..=d {
g.add_edge(0, i, 1, 0, 0);
}
for i in d+1..=2*d {
g.add_edge(i, 2*d+1, 1, 0, 0);
}
for i in 0..d {
for j in 0..d {
if g1[i].is_superset(&g2[j]) {
g.add_edge(1+i, 1+d+j, 1, 0, 0);
}
}
}
g.dinic(0, 2*d+1).0 == d as i64
}
#[inline(never)]
fn add_reduce_maximal<BigNum>(lines : &mut Vec<Line<BigNum>>, newline : Line<BigNum>) where BigNum : crate::bignum::BigNum {
let l1 = lines.len();
lines.retain(|oldline| !perm_includes(&newline, oldline));
let l2 = lines.len();
if l1 != l2 || lines.iter().all(|oldline|!perm_includes(oldline,&newline)) {
lines.push(newline);
}
}
#[inline(never)]
fn find_good_unions<BigNum>(u1 : &[BigNum], u2 : &[BigNum]) -> HashMap<BigNum,Vec<(BigNum,BigNum)>> where BigNum : crate::bignum::BigNum {
let mut unions = HashMap::new();
for x in u1.iter() {
for y in u2.iter() {
if x.is_superset(y) || y.is_superset(x) {
continue;
}
let uni = x.clone() | y.clone();
let unis : &mut Vec<(BigNum,BigNum)> = unions.entry(uni).or_insert(vec![]);
let len = unis.len();
unis.retain(|(xc,yc)| !(xc.is_superset(x) && yc.is_superset(y)) );
if unis.len() != len || unis
|
{
let mut state = vec![0;max.len()];
let mut res = n;
let mut i = 0;
while res > 0 {
let cur = std::cmp::min(max[i],res);
state[i] = cur;
res -= cur;
i += 1;
}
Comb {
max, state, first:true
}
}
|
identifier_body
|
|
forall.rs
|
self.state;
let m = &mut self.max;
let mut i = 0;
loop {
if i == v.len()-1 {
return None;
}
if v[i] > 0 {
v[i+1] += 1;
v[i] -= 1;
if v[i+1] <= m[i+1] {
break;
}
}
i += 1;
}
let mut res = v[0..=i].iter().sum();
let mut j = 0;
while res > 0 {
let cur = std::cmp::min(m[j],res);
v[j] = cur;
res -= cur;
j += 1;
}
for k in j..=i {
v[k] = 0;
}
return Some(&self.state);
}
}
}
struct
|
{
state : Vec<Comb>,
first : bool,
v1 : Vec<usize>
}
impl Matches {
fn new(v1 : Vec<usize>, mut v2 : Vec<usize>) -> Self {
let mut s = vec![];
for &x in &v1 {
let mut c = Comb::new(x,v2.clone());
c.next();
for i in 0..v2.len() {
v2[i] -= c.state[i];
}
s.push(c);
}
Self {
v1, state : s, first : true
}
}
fn next(&mut self) -> Option<&Vec<Comb>> {
if self.first {
self.first = false;
Some(&self.state)
}else {
for i in (0..self.state.len()).rev() {
if self.state[i].next() != None {
for j in i+1..self.state.len() {
let split = self.state.split_at_mut(j);
let p = &split.0[j-1];
let p2 = &mut split.1[0];
let pmax = &p.max;
let ps = &p.state;
let n = self.v1[j];
p2.transform(n,pmax.iter().zip(ps.iter()).map(|(m,x)|m-x));
//let v : Vec<_> = pmax.iter().zip(ps.iter()).map(|(m,x)|m-x).collect();
//self.state[j] = Comb::new(n,v);
p2.next();
}
return Some(&self.state);
}
}
None
}
}
}
fn count_map<BigNum>(v : &[BigNum]) -> HashMap<BigNum,usize> where BigNum : crate::bignum::BigNum{
let mut h = HashMap::new();
for n in v {
*h.entry(n.clone()).or_default() += 1;
}
h
}
#[inline(never)]
fn intersections<BigNum>(uni : BigNum, c1 : &[(BigNum,usize)], c2 : &[(BigNum,usize)], delta : usize , bits : usize) -> Vec<Line<BigNum>> where BigNum : crate::bignum::BigNum {
let v1 : Vec<_> = c1.iter().map(|(_,c)|*c).collect();
let v2 : Vec<_> = c2.iter().map(|(_,c)|*c).collect();
let mut m = Matches::new(v1,v2);
let mut r = vec![];
let mut oldbad : Option<(usize,usize,usize,usize)> = None;
'outer: while let Some(x) = m.next() {
if let Some((i1,i2,j1,j2)) = oldbad {
if x[i1].state[j1] != 0 && x[i2].state[j2] != 0 {
continue 'outer;
}
}
for i1 in 0..c1.len() {
for j1 in 0..c2.len() {
if x[i1].state[j1] != 0 {
for i2 in i1+1..c1.len() {
for j2 in 0..c2.len() {
if x[i2].state[j2] != 0 {
let u1 = c1[i1].0.clone() & c2[j1].0.clone();
let u2 = c1[i2].0.clone() & c2[j2].0.clone();
let u3 = c1[i1].0.clone() & c2[j2].0.clone();
let u4 = c1[i2].0.clone() & c2[j1].0.clone();
if (u4.is_superset(&u1) && u3.is_superset(&u2) && (u1 != u4 || u2 != u3)) || (u3.is_superset(&u1) && u4.is_superset(&u2) && (u1 != u3 || u2 != u4)) {
oldbad = Some((i1,i2,j1,j2));
continue 'outer;
}
}
}
}
}
}
}
let mut groups = Vec::with_capacity(delta);
groups.push(uni.clone());
for (i,(ga,_)) in c1.iter().enumerate() {
for (j,(gb,_)) in c2.iter().enumerate() {
for _ in 0..x[i].state[j] {
groups.push(ga.clone() & gb.clone());
}
}
}
if !groups.contains(&BigNum::zero()) {
r.push(Line::from_groups(delta, bits, groups.into_iter()).sorted());
}
}
r
}
#[inline(never)]
fn perm_includes<BigNum>(line : &Line<BigNum>, other : &Line<BigNum>) -> bool where BigNum : crate::bignum::BigNum {
let g1 : Vec<_> = line.groups().collect();
let g2 : Vec<_> = other.groups().collect();
let d = g1.len();
let mut g = contest_algorithms::graph::flow::FlowGraph::new(2*d+2,d*d);
for i in 1..=d {
g.add_edge(0, i, 1, 0, 0);
}
for i in d+1..=2*d {
g.add_edge(i, 2*d+1, 1, 0, 0);
}
for i in 0..d {
for j in 0..d {
if g1[i].is_superset(&g2[j]) {
g.add_edge(1+i, 1+d+j, 1, 0, 0);
}
}
}
g.dinic(0, 2*d+1).0 == d as i64
}
#[inline(never)]
fn add_reduce_maximal<BigNum>(lines : &mut Vec<Line<BigNum>>, newline : Line<BigNum>) where BigNum : crate::bignum::BigNum {
let l1 = lines.len();
lines.retain(|oldline| !perm_includes(&newline, oldline));
let l2 = lines.len();
if l1 != l2 || lines.iter().all(|oldline|!perm_includes(oldline,&newline)) {
lines.push(newline);
}
}
#[inline(never)]
fn find_good_unions<BigNum>(u1 : &[BigNum], u2 : &[BigNum]) -> HashMap<BigNum,Vec<(BigNum,BigNum)>> where BigNum : crate::bignum::BigNum {
let mut unions = HashMap::new();
for x in u1.iter() {
for y in u2.iter() {
if x.is_superset(y) || y.is_superset(x) {
continue;
}
let uni = x.clone() | y.clone();
let unis : &mut Vec<(BigNum,BigNum)> = unions.entry(uni).or_insert(vec![]);
let len = unis.len();
unis.retain(|(xc,yc)| !(xc.is_superset(x) && yc.is_superset(y)) );
if unis.len() != len || unis.iter().all(|(xc,yc)| !(x.is_superset(xc) && y.is_superset(yc)) ) {
unis.push((x.clone(),y.clone()));
}
}
}
unions
}
pub fn forall<BigNum>(nc : &Constraint<BigNum>, problem : &Problem<BigNum>) -> Constraint<BigNum> where BigNum : crate::bignum::BigNum {
let mut nc = nc.clone();
let maplt = problem.map_label_text();
let set_to_string = |s:&BigNum|{
let r = s.one_bits().map(|elem|&maplt[&elem]).join("");
if r == "" {
String::from("_")
}else{
r
}
};
let make_right_closed = |g : BigNum|{g.clone()|problem.successors(g.one_bits().next().unwrap(),false)};
for line in &mut nc.lines {
*line = line.edited(|g|{make_right_closed(g)}).sorted();
}
let mut seen : HashSet<_> = nc.lines.iter().cloned().collect();
let lines = std::mem::replace(&mut nc.lines, vec![]);
for line in lines {
seen.insert(line.clone());
add
|
Matches
|
identifier_name
|
api_op_CreateFleet.go
|
type CreateFleetInput struct {
// A descriptive label that is associated with a fleet. Fleet names do not need to
// be unique.
//
// This member is required.
Name *string
// Amazon GameLift Anywhere configuration options.
AnywhereConfiguration *types.AnywhereConfiguration
// The unique identifier for a custom game server build to be deployed on fleet
// instances. You can use either the build ID or ARN. The build must be uploaded to
// Amazon GameLift and in READY status. This fleet property cannot be changed
// later.
BuildId *string
// Prompts Amazon GameLift to generate a TLS/SSL certificate for the fleet. Amazon
// GameLift uses the certificates to encrypt traffic between game clients and the
// game servers running on Amazon GameLift. By default, the
// CertificateConfiguration is DISABLED . You can't change this property after you
// create the fleet. Certificate Manager (ACM) certificates expire after 13 months.
// Certificate expiration can cause fleets to fail, preventing players from
// connecting to instances in the fleet. We recommend you replace fleets before 13
// months, consider using fleet aliases for a smooth transition. ACM isn't
// available in all Amazon Web Services regions. A fleet creation request with
// certificate generation enabled in an unsupported Region, fails with a 4xx error.
// For more information about the supported Regions, see Supported Regions (https://docs.aws.amazon.com/acm/latest/userguide/acm-regions.html)
// in the Certificate Manager User Guide.
CertificateConfiguration *types.CertificateConfiguration
// The type of compute resource used to host your game servers. You can use your
// own compute resources with Amazon GameLift Anywhere or use Amazon EC2 instances
// with managed Amazon GameLift. By default, this property is set to EC2 .
ComputeType types.ComputeType
// A description for the fleet.
Description *string
// The allowed IP address ranges and port settings that allow inbound traffic to
// access game sessions on this fleet. If the fleet is hosting a custom game build,
// this property must be set before players can connect to game sessions. For
// Realtime Servers fleets, Amazon GameLift automatically sets TCP and UDP ranges.
EC2InboundPermissions []types.IpPermission
// The Amazon GameLift-supported Amazon EC2 instance type to use for all fleet
// instances. Instance type determines the computing resources that will be used to
// host your game servers, including CPU, memory, storage, and networking capacity.
// See Amazon Elastic Compute Cloud Instance Types (http://aws.amazon.com/ec2/instance-types/)
// for detailed descriptions of Amazon EC2 instance types.
EC2InstanceType types.EC2InstanceType
// Indicates whether to use On-Demand or Spot instances for this fleet. By
// default, this property is set to ON_DEMAND . Learn more about when to use
// On-Demand versus Spot Instances (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-ec2-instances.html#gamelift-ec2-instances-spot)
// . This property cannot be changed after the fleet is created.
FleetType types.FleetType
// A unique identifier for an IAM role that manages access to your Amazon Web
// Services services. With an instance role ARN set, any application that runs on
// an instance in this fleet can assume the role, including install scripts, server
// processes, and daemons (background processes). Create a role or look up a role's
// ARN by using the IAM dashboard (https://console.aws.amazon.com/iam/) in the
// Amazon Web Services Management Console. Learn more about using on-box
// credentials for your game servers at Access external resources from a game
// server (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-resources.html)
// . This property cannot be changed after the fleet is created.
InstanceRoleArn *string
// A set of remote locations to deploy additional instances to and manage as part
// of the fleet. This parameter can only be used when creating fleets in Amazon Web
// Services Regions that support multiple locations. You can add any Amazon
// GameLift-supported Amazon Web Services Region as a remote location, in the form
// of an Amazon Web Services Region code such as us-west-2 . To create a fleet with
// instances in the home Region only, don't use this parameter. To use this
// parameter, Amazon GameLift requires you to use your home location in the
// request.
Locations []types.LocationConfiguration
// This parameter is no longer used. To specify where Amazon GameLift should store
// log files once a server process shuts down, use the Amazon GameLift server API
// ProcessReady() and specify one or more directory paths in logParameters . For
// more information, see Initialize the server process (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-api.html#gamelift-sdk-server-initialize)
// in the Amazon GameLift Developer Guide.
LogPaths []string
// The name of an Amazon Web Services CloudWatch metric group to add this fleet
// to. A metric group is used to aggregate the metrics for multiple fleets. You can
// specify an existing metric group name or set a new name to create a new metric
// group. A fleet can be included in only one metric group at a time.
MetricGroups []string
// The status of termination protection for active game sessions on the fleet. By
// default, this property is set to NoProtection . You can also set game session
// protection for an individual game session by calling UpdateGameSession .
// - NoProtection - Game sessions can be terminated during active gameplay as a
// result of a scale-down event.
// - FullProtection - Game sessions in ACTIVE status cannot be terminated during
// a scale-down event.
NewGameSessionProtectionPolicy types.ProtectionPolicy
// Used when peering your Amazon GameLift fleet with a VPC, the unique identifier
// for the Amazon Web Services account that owns the VPC. You can find your account
// ID in the Amazon Web Services Management Console under account settings.
PeerVpcAwsAccountId *string
// A unique identifier for a VPC with resources to be accessed by your Amazon
// GameLift fleet. The VPC must be in the same Region as your fleet. To look up a
// VPC ID, use the VPC Dashboard (https://console.aws.amazon.com/vpc/) in the
// Amazon Web Services Management Console. Learn more about VPC peering in VPC
// Peering with Amazon GameLift Fleets (https://docs.aws.amazon.com/gamelift/latest/developerguide/vpc-peering.html)
// .
PeerVpcId *string
// A policy that limits the number of game sessions that an individual player can
// create on instances in this fleet within a specified span of time.
ResourceCreationLimitPolicy *types.ResourceCreationLimitPolicy
// Instructions for how to launch and maintain server processes on instances in
// the fleet. The runtime configuration defines one or more server process
// configurations, each identifying a build executable or Realtime script file and
// the number of processes of that type to run concurrently. The
// RuntimeConfiguration parameter is required unless the fleet is being configured
// using the older parameters ServerLaunchPath and ServerLaunchParameters , which
// are still supported for backward compatibility.
RuntimeConfiguration *types.RuntimeConfiguration
// The unique identifier for a Realtime configuration script to be deployed on
// fleet instances. You can use either the script ID or ARN. Scripts must be
// uploaded to Amazon GameLift prior to creating the fleet. This fleet property
// cannot be changed later.
ScriptId *string
// This parameter is no longer used. Specify server launch parameters using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchParameters *string
// This parameter is no longer used. Specify a server launch path using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchPath *string
// A list of labels to assign to the new fleet resource. Tags are
// developer-defined key-value pairs. Tagging Amazon Web Services resources are
// useful for resource management, access management and cost allocation. For more
// information, see Tagging Amazon Web Services Resources (https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html)
// in the Amazon Web Services General Reference.
Tags []types.Tag
noSmithyDocumentSerde
}
type CreateFleetOutput struct {
// The properties for the new fleet, including the current status. All fleets are
// placed in NEW status on creation.
|
{
if params == nil {
params = &CreateFleetInput{}
}
result, metadata, err := c.invokeOperation(ctx, "CreateFleet", params, optFns, c.addOperationCreateFleetMiddlewares)
if err != nil {
return nil, err
}
out := result.(*CreateFleetOutput)
out.ResultMetadata = metadata
return out, nil
}
|
identifier_body
|
|
api_op_CreateFleet.go
|
// computing resources for your fleet and provide instructions for running game
// servers on each instance. Most Amazon GameLift fleets can deploy instances to
// multiple locations, including the home Region (where the fleet is created) and
// an optional set of remote locations. Fleets that are created in the following
// Amazon Web Services Regions support multiple locations: us-east-1 (N. Virginia),
// us-west-2 (Oregon), eu-central-1 (Frankfurt), eu-west-1 (Ireland),
// ap-southeast-2 (Sydney), ap-northeast-1 (Tokyo), and ap-northeast-2 (Seoul).
// Fleets that are created in other Amazon GameLift Regions can deploy instances in
// the fleet's home Region only. All fleet instances use the same configuration
// regardless of location; however, you can adjust capacity settings and turn
// auto-scaling on/off for each location. To create a fleet, choose the hardware
// for your instances, specify a game server build or Realtime script to deploy,
// and provide a runtime configuration to direct Amazon GameLift how to start and
// run game servers on each instance in the fleet. Set permissions for inbound
// traffic to your game servers, and enable optional features as needed. When
// creating a multi-location fleet, provide a list of additional remote locations.
// If you need to debug your fleet, fetch logs, view performance metrics or other
// actions on the fleet, create the development fleet with port 22/3389 open. As a
// best practice, we recommend opening ports for remote access only when you need
// them and closing them when you're finished. If successful, this operation
// creates a new Fleet resource and places it in NEW status, which prompts Amazon
// GameLift to initiate the fleet creation workflow (https://docs.aws.amazon.com/gamelift/latest/developerguide/fleets-creating-all.html#fleets-creation-workflow)
// . Learn more Setting up fleets (https://docs.aws.amazon.com/gamelift/latest/developerguide/fleets-intro.html)
// Debug fleet creation issues (https://docs.aws.amazon.com/gamelift/latest/developerguide/fleets-creating-debug.html#fleets-creating-debug-creation)
// Multi-location fleets (https://docs.aws.amazon.com/gamelift/latest/developerguide/fleets-intro.html)
func (c *Client) CreateFleet(ctx context.Context, params *CreateFleetInput, optFns ...func(*Options)) (*CreateFleetOutput, error) {
if params == nil {
params = &CreateFleetInput{}
}
result, metadata, err := c.invokeOperation(ctx, "CreateFleet", params, optFns, c.addOperationCreateFleetMiddlewares)
if err != nil {
return nil, err
}
out := result.(*CreateFleetOutput)
out.ResultMetadata = metadata
return out, nil
}
type CreateFleetInput struct {
// A descriptive label that is associated with a fleet. Fleet names do not need to
// be unique.
//
// This member is required.
Name *string
// Amazon GameLift Anywhere configuration options.
AnywhereConfiguration *types.AnywhereConfiguration
// The unique identifier for a custom game server build to be deployed on fleet
// instances. You can use either the build ID or ARN. The build must be uploaded to
// Amazon GameLift and in READY status. This fleet property cannot be changed
// later.
BuildId *string
// Prompts Amazon GameLift to generate a TLS/SSL certificate for the fleet. Amazon
// GameLift uses the certificates to encrypt traffic between game clients and the
// game servers running on Amazon GameLift. By default, the
// CertificateConfiguration is DISABLED . You can't change this property after you
// create the fleet. Certificate Manager (ACM) certificates expire after 13 months.
// Certificate expiration can cause fleets to fail, preventing players from
// connecting to instances in the fleet. We recommend you replace fleets before 13
// months, consider using fleet aliases for a smooth transition. ACM isn't
// available in all Amazon Web Services regions. A fleet creation request with
// certificate generation enabled in an unsupported Region, fails with a 4xx error.
// For more information about the supported Regions, see Supported Regions (https://docs.aws.amazon.com/acm/latest/userguide/acm-regions.html)
// in the Certificate Manager User Guide.
CertificateConfiguration *types.CertificateConfiguration
// The type of compute resource used to host your game servers. You can use your
// own compute resources with Amazon GameLift Anywhere or use Amazon EC2 instances
// with managed Amazon GameLift. By default, this property is set to EC2 .
ComputeType types.ComputeType
// A description for the fleet.
Description *string
// The allowed IP address ranges and port settings that allow inbound traffic to
// access game sessions on this fleet. If the fleet is hosting a custom game build,
// this property must be set before players can connect to game sessions. For
// Realtime Servers fleets, Amazon GameLift automatically sets TCP and UDP ranges.
EC2InboundPermissions []types.IpPermission
// The Amazon GameLift-supported Amazon EC2 instance type to use for all fleet
// instances. Instance type determines the computing resources that will be used to
// host your game servers, including CPU, memory, storage, and networking capacity.
// See Amazon Elastic Compute Cloud Instance Types (http://aws.amazon.com/ec2/instance-types/)
// for detailed descriptions of Amazon EC2 instance types.
EC2InstanceType types.EC2InstanceType
// Indicates whether to use On-Demand or Spot instances for this fleet. By
// default, this property is set to ON_DEMAND . Learn more about when to use
// On-Demand versus Spot Instances (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-ec2-instances.html#gamelift-ec2-instances-spot)
// . This property cannot be changed after the fleet is created.
FleetType types.FleetType
// A unique identifier for an IAM role that manages access to your Amazon Web
// Services services. With an instance role ARN set, any application that runs on
// an instance in this fleet can assume the role, including install scripts, server
// processes, and daemons (background processes). Create a role or look up a role's
// ARN by using the IAM dashboard (https://console.aws.amazon.com/iam/) in the
// Amazon Web Services Management Console. Learn more about using on-box
// credentials for your game servers at Access external resources from a game
// server (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-resources.html)
// . This property cannot be changed after the fleet is created.
InstanceRoleArn *string
// A set of remote locations to deploy additional instances to and manage as part
// of the fleet. This parameter can only be used when creating fleets in Amazon Web
// Services Regions that support multiple locations. You can add any Amazon
// GameLift-supported Amazon Web Services Region as a remote location, in the form
// of an Amazon Web Services Region code such as us-west-2 . To create a fleet with
// instances in the home Region only, don't use this parameter. To use this
// parameter, Amazon GameLift requires you to use your home location in the
// request.
Locations []types.LocationConfiguration
// This parameter is no longer used. To specify where Amazon GameLift should store
// log files once a server process shuts down, use the Amazon GameLift server API
// ProcessReady() and specify one or more directory paths in logParameters . For
// more information, see Initialize the server process (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-api.html#gamelift-sdk-server-initialize)
// in the Amazon GameLift Developer Guide.
LogPaths []string
// The name of an Amazon Web Services CloudWatch metric group to add this fleet
// to. A metric group is used to aggregate the metrics for multiple fleets. You can
// specify an existing metric group name or set a new name to create a new metric
// group. A fleet can be included in only one metric group at a time.
MetricGroups []string
// The status of termination protection for active game sessions on the fleet. By
// default, this property is set to NoProtection . You can also set game session
// protection for an individual game session by calling UpdateGameSession .
// - NoProtection - Game sessions can be terminated during active gameplay as a
// result of a scale-down event.
// - FullProtection - Game sessions in ACTIVE status cannot be terminated during
// a scale-down event.
NewGameSessionProtectionPolicy types.ProtectionPolicy
// Used when peering your Amazon GameLift fleet with a VPC, the unique identifier
// for the Amazon Web Services account that owns the VPC. You can find your account
// ID in the
|
)
// Creates a fleet of Amazon Elastic Compute Cloud (Amazon EC2) instances to host
// your custom game server or Realtime Servers. Use this operation to configure the
|
random_line_split
|
|
api_op_CreateFleet.go
|
READY status. This fleet property cannot be changed
// later.
BuildId *string
// Prompts Amazon GameLift to generate a TLS/SSL certificate for the fleet. Amazon
// GameLift uses the certificates to encrypt traffic between game clients and the
// game servers running on Amazon GameLift. By default, the
// CertificateConfiguration is DISABLED . You can't change this property after you
// create the fleet. Certificate Manager (ACM) certificates expire after 13 months.
// Certificate expiration can cause fleets to fail, preventing players from
// connecting to instances in the fleet. We recommend you replace fleets before 13
// months, consider using fleet aliases for a smooth transition. ACM isn't
// available in all Amazon Web Services regions. A fleet creation request with
// certificate generation enabled in an unsupported Region, fails with a 4xx error.
// For more information about the supported Regions, see Supported Regions (https://docs.aws.amazon.com/acm/latest/userguide/acm-regions.html)
// in the Certificate Manager User Guide.
CertificateConfiguration *types.CertificateConfiguration
// The type of compute resource used to host your game servers. You can use your
// own compute resources with Amazon GameLift Anywhere or use Amazon EC2 instances
// with managed Amazon GameLift. By default, this property is set to EC2 .
ComputeType types.ComputeType
// A description for the fleet.
Description *string
// The allowed IP address ranges and port settings that allow inbound traffic to
// access game sessions on this fleet. If the fleet is hosting a custom game build,
// this property must be set before players can connect to game sessions. For
// Realtime Servers fleets, Amazon GameLift automatically sets TCP and UDP ranges.
EC2InboundPermissions []types.IpPermission
// The Amazon GameLift-supported Amazon EC2 instance type to use for all fleet
// instances. Instance type determines the computing resources that will be used to
// host your game servers, including CPU, memory, storage, and networking capacity.
// See Amazon Elastic Compute Cloud Instance Types (http://aws.amazon.com/ec2/instance-types/)
// for detailed descriptions of Amazon EC2 instance types.
EC2InstanceType types.EC2InstanceType
// Indicates whether to use On-Demand or Spot instances for this fleet. By
// default, this property is set to ON_DEMAND . Learn more about when to use
// On-Demand versus Spot Instances (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-ec2-instances.html#gamelift-ec2-instances-spot)
// . This property cannot be changed after the fleet is created.
FleetType types.FleetType
// A unique identifier for an IAM role that manages access to your Amazon Web
// Services services. With an instance role ARN set, any application that runs on
// an instance in this fleet can assume the role, including install scripts, server
// processes, and daemons (background processes). Create a role or look up a role's
// ARN by using the IAM dashboard (https://console.aws.amazon.com/iam/) in the
// Amazon Web Services Management Console. Learn more about using on-box
// credentials for your game servers at Access external resources from a game
// server (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-resources.html)
// . This property cannot be changed after the fleet is created.
InstanceRoleArn *string
// A set of remote locations to deploy additional instances to and manage as part
// of the fleet. This parameter can only be used when creating fleets in Amazon Web
// Services Regions that support multiple locations. You can add any Amazon
// GameLift-supported Amazon Web Services Region as a remote location, in the form
// of an Amazon Web Services Region code such as us-west-2 . To create a fleet with
// instances in the home Region only, don't use this parameter. To use this
// parameter, Amazon GameLift requires you to use your home location in the
// request.
Locations []types.LocationConfiguration
// This parameter is no longer used. To specify where Amazon GameLift should store
// log files once a server process shuts down, use the Amazon GameLift server API
// ProcessReady() and specify one or more directory paths in logParameters . For
// more information, see Initialize the server process (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-api.html#gamelift-sdk-server-initialize)
// in the Amazon GameLift Developer Guide.
LogPaths []string
// The name of an Amazon Web Services CloudWatch metric group to add this fleet
// to. A metric group is used to aggregate the metrics for multiple fleets. You can
// specify an existing metric group name or set a new name to create a new metric
// group. A fleet can be included in only one metric group at a time.
MetricGroups []string
// The status of termination protection for active game sessions on the fleet. By
// default, this property is set to NoProtection . You can also set game session
// protection for an individual game session by calling UpdateGameSession .
// - NoProtection - Game sessions can be terminated during active gameplay as a
// result of a scale-down event.
// - FullProtection - Game sessions in ACTIVE status cannot be terminated during
// a scale-down event.
NewGameSessionProtectionPolicy types.ProtectionPolicy
// Used when peering your Amazon GameLift fleet with a VPC, the unique identifier
// for the Amazon Web Services account that owns the VPC. You can find your account
// ID in the Amazon Web Services Management Console under account settings.
PeerVpcAwsAccountId *string
// A unique identifier for a VPC with resources to be accessed by your Amazon
// GameLift fleet. The VPC must be in the same Region as your fleet. To look up a
// VPC ID, use the VPC Dashboard (https://console.aws.amazon.com/vpc/) in the
// Amazon Web Services Management Console. Learn more about VPC peering in VPC
// Peering with Amazon GameLift Fleets (https://docs.aws.amazon.com/gamelift/latest/developerguide/vpc-peering.html)
// .
PeerVpcId *string
// A policy that limits the number of game sessions that an individual player can
// create on instances in this fleet within a specified span of time.
ResourceCreationLimitPolicy *types.ResourceCreationLimitPolicy
// Instructions for how to launch and maintain server processes on instances in
// the fleet. The runtime configuration defines one or more server process
// configurations, each identifying a build executable or Realtime script file and
// the number of processes of that type to run concurrently. The
// RuntimeConfiguration parameter is required unless the fleet is being configured
// using the older parameters ServerLaunchPath and ServerLaunchParameters , which
// are still supported for backward compatibility.
RuntimeConfiguration *types.RuntimeConfiguration
// The unique identifier for a Realtime configuration script to be deployed on
// fleet instances. You can use either the script ID or ARN. Scripts must be
// uploaded to Amazon GameLift prior to creating the fleet. This fleet property
// cannot be changed later.
ScriptId *string
// This parameter is no longer used. Specify server launch parameters using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchParameters *string
// This parameter is no longer used. Specify a server launch path using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchPath *string
// A list of labels to assign to the new fleet resource. Tags are
// developer-defined key-value pairs. Tagging Amazon Web Services resources are
// useful for resource management, access management and cost allocation. For more
// information, see Tagging Amazon Web Services Resources (https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html)
// in the Amazon Web Services General Reference.
Tags []types.Tag
noSmithyDocumentSerde
}
type CreateFleetOutput struct {
// The properties for the new fleet, including the current status. All fleets are
// placed in NEW status on creation.
FleetAttributes *types.FleetAttributes
// The fleet's locations and life-cycle status of each location. For new fleets,
// the status of all locations is set to NEW . During fleet creation, Amazon
// GameLift updates each location status as instances are deployed there and
// prepared for game hosting. This list includes an entry for the fleet's home
// Region. For fleets with no remote locations, only one entry, representing the
// home Region, is returned.
LocationStates []types.LocationState
// Metadata pertaining to the operation's result.
ResultMetadata middleware.Metadata
noSmithyDocumentSerde
}
func (c *Client) addOperationCreateFleetMiddlewares(stack *middleware.Stack, options Options) (err error) {
err = stack.Serialize.Add(&awsAwsjson11_serializeOpCreateFleet{}, middleware.After)
if err != nil
|
{
return err
}
|
conditional_block
|
|
api_op_CreateFleet.go
|
(ctx context.Context, params *CreateFleetInput, optFns ...func(*Options)) (*CreateFleetOutput, error) {
if params == nil {
params = &CreateFleetInput{}
}
result, metadata, err := c.invokeOperation(ctx, "CreateFleet", params, optFns, c.addOperationCreateFleetMiddlewares)
if err != nil {
return nil, err
}
out := result.(*CreateFleetOutput)
out.ResultMetadata = metadata
return out, nil
}
type CreateFleetInput struct {
// A descriptive label that is associated with a fleet. Fleet names do not need to
// be unique.
//
// This member is required.
Name *string
// Amazon GameLift Anywhere configuration options.
AnywhereConfiguration *types.AnywhereConfiguration
// The unique identifier for a custom game server build to be deployed on fleet
// instances. You can use either the build ID or ARN. The build must be uploaded to
// Amazon GameLift and in READY status. This fleet property cannot be changed
// later.
BuildId *string
// Prompts Amazon GameLift to generate a TLS/SSL certificate for the fleet. Amazon
// GameLift uses the certificates to encrypt traffic between game clients and the
// game servers running on Amazon GameLift. By default, the
// CertificateConfiguration is DISABLED . You can't change this property after you
// create the fleet. Certificate Manager (ACM) certificates expire after 13 months.
// Certificate expiration can cause fleets to fail, preventing players from
// connecting to instances in the fleet. We recommend you replace fleets before 13
// months, consider using fleet aliases for a smooth transition. ACM isn't
// available in all Amazon Web Services regions. A fleet creation request with
// certificate generation enabled in an unsupported Region, fails with a 4xx error.
// For more information about the supported Regions, see Supported Regions (https://docs.aws.amazon.com/acm/latest/userguide/acm-regions.html)
// in the Certificate Manager User Guide.
CertificateConfiguration *types.CertificateConfiguration
// The type of compute resource used to host your game servers. You can use your
// own compute resources with Amazon GameLift Anywhere or use Amazon EC2 instances
// with managed Amazon GameLift. By default, this property is set to EC2 .
ComputeType types.ComputeType
// A description for the fleet.
Description *string
// The allowed IP address ranges and port settings that allow inbound traffic to
// access game sessions on this fleet. If the fleet is hosting a custom game build,
// this property must be set before players can connect to game sessions. For
// Realtime Servers fleets, Amazon GameLift automatically sets TCP and UDP ranges.
EC2InboundPermissions []types.IpPermission
// The Amazon GameLift-supported Amazon EC2 instance type to use for all fleet
// instances. Instance type determines the computing resources that will be used to
// host your game servers, including CPU, memory, storage, and networking capacity.
// See Amazon Elastic Compute Cloud Instance Types (http://aws.amazon.com/ec2/instance-types/)
// for detailed descriptions of Amazon EC2 instance types.
EC2InstanceType types.EC2InstanceType
// Indicates whether to use On-Demand or Spot instances for this fleet. By
// default, this property is set to ON_DEMAND . Learn more about when to use
// On-Demand versus Spot Instances (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-ec2-instances.html#gamelift-ec2-instances-spot)
// . This property cannot be changed after the fleet is created.
FleetType types.FleetType
// A unique identifier for an IAM role that manages access to your Amazon Web
// Services services. With an instance role ARN set, any application that runs on
// an instance in this fleet can assume the role, including install scripts, server
// processes, and daemons (background processes). Create a role or look up a role's
// ARN by using the IAM dashboard (https://console.aws.amazon.com/iam/) in the
// Amazon Web Services Management Console. Learn more about using on-box
// credentials for your game servers at Access external resources from a game
// server (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-resources.html)
// . This property cannot be changed after the fleet is created.
InstanceRoleArn *string
// A set of remote locations to deploy additional instances to and manage as part
// of the fleet. This parameter can only be used when creating fleets in Amazon Web
// Services Regions that support multiple locations. You can add any Amazon
// GameLift-supported Amazon Web Services Region as a remote location, in the form
// of an Amazon Web Services Region code such as us-west-2 . To create a fleet with
// instances in the home Region only, don't use this parameter. To use this
// parameter, Amazon GameLift requires you to use your home location in the
// request.
Locations []types.LocationConfiguration
// This parameter is no longer used. To specify where Amazon GameLift should store
// log files once a server process shuts down, use the Amazon GameLift server API
// ProcessReady() and specify one or more directory paths in logParameters . For
// more information, see Initialize the server process (https://docs.aws.amazon.com/gamelift/latest/developerguide/gamelift-sdk-server-api.html#gamelift-sdk-server-initialize)
// in the Amazon GameLift Developer Guide.
LogPaths []string
// The name of an Amazon Web Services CloudWatch metric group to add this fleet
// to. A metric group is used to aggregate the metrics for multiple fleets. You can
// specify an existing metric group name or set a new name to create a new metric
// group. A fleet can be included in only one metric group at a time.
MetricGroups []string
// The status of termination protection for active game sessions on the fleet. By
// default, this property is set to NoProtection . You can also set game session
// protection for an individual game session by calling UpdateGameSession .
// - NoProtection - Game sessions can be terminated during active gameplay as a
// result of a scale-down event.
// - FullProtection - Game sessions in ACTIVE status cannot be terminated during
// a scale-down event.
NewGameSessionProtectionPolicy types.ProtectionPolicy
// Used when peering your Amazon GameLift fleet with a VPC, the unique identifier
// for the Amazon Web Services account that owns the VPC. You can find your account
// ID in the Amazon Web Services Management Console under account settings.
PeerVpcAwsAccountId *string
// A unique identifier for a VPC with resources to be accessed by your Amazon
// GameLift fleet. The VPC must be in the same Region as your fleet. To look up a
// VPC ID, use the VPC Dashboard (https://console.aws.amazon.com/vpc/) in the
// Amazon Web Services Management Console. Learn more about VPC peering in VPC
// Peering with Amazon GameLift Fleets (https://docs.aws.amazon.com/gamelift/latest/developerguide/vpc-peering.html)
// .
PeerVpcId *string
// A policy that limits the number of game sessions that an individual player can
// create on instances in this fleet within a specified span of time.
ResourceCreationLimitPolicy *types.ResourceCreationLimitPolicy
// Instructions for how to launch and maintain server processes on instances in
// the fleet. The runtime configuration defines one or more server process
// configurations, each identifying a build executable or Realtime script file and
// the number of processes of that type to run concurrently. The
// RuntimeConfiguration parameter is required unless the fleet is being configured
// using the older parameters ServerLaunchPath and ServerLaunchParameters , which
// are still supported for backward compatibility.
RuntimeConfiguration *types.RuntimeConfiguration
// The unique identifier for a Realtime configuration script to be deployed on
// fleet instances. You can use either the script ID or ARN. Scripts must be
// uploaded to Amazon GameLift prior to creating the fleet. This fleet property
// cannot be changed later.
ScriptId *string
// This parameter is no longer used. Specify server launch parameters using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchParameters *string
// This parameter is no longer used. Specify a server launch path using the
// RuntimeConfiguration parameter. Requests that use this parameter instead
// continue to be valid.
ServerLaunchPath *string
// A list of labels to assign to the new fleet resource. Tags are
// developer-defined key-value pairs. Tagging Amazon Web Services resources are
// useful for resource management, access management and cost allocation. For more
// information, see Tagging Amazon Web Services Resources (https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html)
// in the Amazon Web Services General Reference.
Tags []types.Tag
noSmithyDocumentSerde
}
type CreateFleetOutput
|
CreateFleet
|
identifier_name
|
|
policy_handler.rs
|
]
pub trait PolicyHandler {
/// Called when a policy client makes a request on the policy API this handler controls.
async fn handle_policy_request(&mut self, request: PolicyRequest) -> Response;
/// Called when a setting request is intercepted for the setting this policy handler supervises.
///
/// If there are no policies or the request does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to consume the request and respond to the client directly, it should
/// return [`RequestTransform::Result`].
///
/// If this handler wants to modify the request, then let the setting handler handle it,
/// [`RequestTransform::Request`] should be returned, with the modified request.
///
/// [`RequestTransform::Result`]: enum.RequestTransform.html
/// [`RequestTransform::Request`]: enum.RequestTransform.html
async fn handle_setting_request(&mut self, request: Request) -> Option<RequestTransform>;
/// Called when a setting response is intercepted from the setting this policy handler
/// supervises.
///
/// If there are no policies or the response does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to modify the response and still let the original audience handle it,
/// [`Response`] should be returned, containing the modified response.
///
/// [`Response`]: ResponseTransform::Response
async fn handle_setting_response(
&mut self,
response: SettingResponse,
) -> Option<ResponseTransform>;
}
/// `RequestTransform` is returned by a [`PolicyHandler`] in response to a setting request that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum RequestTransform {
/// A new, modified request that should be forwarded to the setting handler for processing.
Request(Request),
/// A result to return directly to the settings client.
Result(SettingHandlerResult),
}
/// `ResponseTransform` is returned by a [`PolicyHandler`] in response to a setting response that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum ResponseTransform {
/// A new, modified response that should be forwarded.
Response(SettingResponse),
}
/// Trait used to create policy handlers.
#[async_trait]
pub trait Create: Sized {
async fn create(handler: ClientProxy) -> Result<Self, Error>;
}
/// Creates a [`PolicyHandler`] from the given [`Context`].
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`Context`]: ../base/struct.Context.html
pub(crate) fn create_handler<C, T: StorageFactory<Storage = DeviceStorage> + 'static>(
context: Context<T>,
) -> BoxFuture<'static, GenerateHandlerResult>
where
C: Create + PolicyHandler + Send + Sync + 'static,
{
Box::pin(async move {
let _ = &context;
let proxy = ClientProxy::new(context.service_messenger);
C::create(proxy).await.map(|handler| Box::new(handler) as BoxedHandler)
})
}
/// `ClientProxy` provides common functionality, like messaging and persistence to policy handlers.
#[derive(Clone)]
pub struct ClientProxy {
service_messenger: service::message::Messenger,
}
impl ClientProxy {
/// Sends a setting request to the underlying setting proxy this policy handler controls.
pub(crate) fn send_setting_request(
&self,
setting_type: SettingType,
request: Request,
) -> service::message::Receptor {
self.service_messenger
.message(
HandlerPayload::Request(request).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send()
}
/// Requests the setting handler to rebroadcast a settings changed event to its listeners.
pub(crate) fn request_rebroadcast(&self, setting_type: SettingType) {
// Ignore the receptor result.
let _ = self
.service_messenger
.message(
HandlerPayload::Request(Request::Rebroadcast).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send();
}
}
impl ClientProxy {
pub(crate) fn new(service_messenger: service::message::Messenger) -> Self {
Self { service_messenger }
}
/// The type `T` is any type that has a [`PolicyType`] associated with it and that can be
/// converted into a [`PolicyInfo`]. This is usually a variant of the `PolicyInfo` enum.
pub(crate) async fn read_policy<T: HasPolicyType + TryFrom<PolicyInfo>>(
&self,
id: ftrace::Id,
) -> T {
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Read(T::POLICY_TYPE.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Read(
StorageInfo::PolicyInfo(policy_info),
)) = payload
{
let policy_type: PolicyType = (&policy_info).into();
if let Ok(info) = policy_info.try_into() {
return info;
}
panic!(
"Mismatching type during read. Expected {:?}, but got {:?}",
T::POLICY_TYPE,
policy_type
);
} else {
panic!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
panic!("Error reading from storage: {:?}", err);
}
}
}
/// Write a policy info object to storage. The argument `write_through` will
/// block returning until the value has been completely written to
/// persistent store, rather than any temporary in-memory caching.
pub(crate) async fn write_policy(
&self,
policy_info: PolicyInfo,
id: ftrace::Id,
) -> Result<UpdateState, PolicyError> {
let policy_type = (&policy_info).into();
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Write(policy_info.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Write(result)) = payload
{
return result.map_err(|e| {
fx_log_err!("Failed to write policy: {:?}", e);
PolicyError::WriteFailure(policy_type)
});
} else {
fx_log_err!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
fx_log_err!("Error writing to storage: {:?}", err);
}
}
Err(PolicyError::WriteFailure(policy_type))
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::base::SettingType;
use crate::handler::base::{Payload as HandlerPayload, Request};
use crate::message::base::MessengerType;
use crate::message::MessageHubUtil;
use crate::policy::PolicyType;
use crate::service;
use crate::tests::message_utils::verify_payload;
#[fuchsia_async::run_until_stalled(test)]
async fn test_client_proxy_send_setting_request() {
let policy_type = PolicyType::Unknown;
let setting_request = Request::Get;
let target_setting_type = SettingType::Unknown;
let service_delegate = service::MessageHub::create_hub();
let (_, mut setting_proxy_receptor) = service_delegate
.create(MessengerType::Addressable(service::Address::Handler(
policy_type.setting_type(),
)))
.await
.expect("setting proxy messenger created");
let client_proxy = ClientProxy {
service_messenger: service_delegate
.create(MessengerType::Unbound)
.await
.expect("messenger should be created")
.0,
};
let _ = client_proxy.send_setting_request(target_setting_type, setting_request.clone());
verify_payload(
service::Payload::Setting(HandlerPayload::Request(setting_request)),
&mut setting_proxy_receptor,
None,
)
.await
}
#[fuchsia_async::run_until_stalled(test)]
async fn test_client_proxy_request_rebroadcast()
|
{
let setting_type = SettingType::Unknown;
let service_delegate = service::MessageHub::create_hub();
let (_, mut receptor) = service_delegate
.create(MessengerType::Addressable(service::Address::Handler(setting_type)))
.await
.expect("service receptor created");
let client_proxy = ClientProxy {
service_messenger: service_delegate
.create(MessengerType::Unbound)
.await
.expect("messenger should be created")
.0,
};
client_proxy.request_rebroadcast(setting_type);
|
identifier_body
|
|
policy_handler.rs
|
Factory};
use crate::message::base::Audience;
use crate::policy::response::{Error as PolicyError, Response};
use crate::policy::{
BoxedHandler, Context, GenerateHandlerResult, HasPolicyType, PolicyInfo, PolicyType,
Request as PolicyRequest,
};
use crate::service;
use crate::storage::{self, StorageInfo};
use anyhow::Error;
use async_trait::async_trait;
use fuchsia_syslog::fx_log_err;
use fuchsia_trace as ftrace;
use futures::future::BoxFuture;
use settings_storage::device_storage::DeviceStorage;
use settings_storage::UpdateState;
use std::convert::{TryFrom, TryInto};
/// PolicyHandlers are in charge of applying and persisting policies set by clients.
#[async_trait]
pub trait PolicyHandler {
/// Called when a policy client makes a request on the policy API this handler controls.
async fn handle_policy_request(&mut self, request: PolicyRequest) -> Response;
/// Called when a setting request is intercepted for the setting this policy handler supervises.
///
/// If there are no policies or the request does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to consume the request and respond to the client directly, it should
/// return [`RequestTransform::Result`].
///
/// If this handler wants to modify the request, then let the setting handler handle it,
/// [`RequestTransform::Request`] should be returned, with the modified request.
///
/// [`RequestTransform::Result`]: enum.RequestTransform.html
/// [`RequestTransform::Request`]: enum.RequestTransform.html
async fn handle_setting_request(&mut self, request: Request) -> Option<RequestTransform>;
/// Called when a setting response is intercepted from the setting this policy handler
/// supervises.
///
/// If there are no policies or the response does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to modify the response and still let the original audience handle it,
/// [`Response`] should be returned, containing the modified response.
///
/// [`Response`]: ResponseTransform::Response
async fn handle_setting_response(
&mut self,
response: SettingResponse,
) -> Option<ResponseTransform>;
}
/// `RequestTransform` is returned by a [`PolicyHandler`] in response to a setting request that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum RequestTransform {
/// A new, modified request that should be forwarded to the setting handler for processing.
Request(Request),
/// A result to return directly to the settings client.
Result(SettingHandlerResult),
}
/// `ResponseTransform` is returned by a [`PolicyHandler`] in response to a setting response that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum ResponseTransform {
/// A new, modified response that should be forwarded.
Response(SettingResponse),
}
/// Trait used to create policy handlers.
#[async_trait]
pub trait Create: Sized {
async fn create(handler: ClientProxy) -> Result<Self, Error>;
}
/// Creates a [`PolicyHandler`] from the given [`Context`].
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`Context`]: ../base/struct.Context.html
pub(crate) fn create_handler<C, T: StorageFactory<Storage = DeviceStorage> + 'static>(
context: Context<T>,
) -> BoxFuture<'static, GenerateHandlerResult>
where
C: Create + PolicyHandler + Send + Sync + 'static,
{
Box::pin(async move {
let _ = &context;
let proxy = ClientProxy::new(context.service_messenger);
C::create(proxy).await.map(|handler| Box::new(handler) as BoxedHandler)
})
}
/// `ClientProxy` provides common functionality, like messaging and persistence to policy handlers.
#[derive(Clone)]
pub struct ClientProxy {
service_messenger: service::message::Messenger,
}
impl ClientProxy {
/// Sends a setting request to the underlying setting proxy this policy handler controls.
pub(crate) fn send_setting_request(
&self,
setting_type: SettingType,
request: Request,
) -> service::message::Receptor {
self.service_messenger
.message(
HandlerPayload::Request(request).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send()
}
/// Requests the setting handler to rebroadcast a settings changed event to its listeners.
pub(crate) fn request_rebroadcast(&self, setting_type: SettingType) {
// Ignore the receptor result.
let _ = self
.service_messenger
.message(
HandlerPayload::Request(Request::Rebroadcast).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send();
}
}
impl ClientProxy {
pub(crate) fn new(service_messenger: service::message::Messenger) -> Self {
Self { service_messenger }
}
/// The type `T` is any type that has a [`PolicyType`] associated with it and that can be
/// converted into a [`PolicyInfo`]. This is usually a variant of the `PolicyInfo` enum.
pub(crate) async fn read_policy<T: HasPolicyType + TryFrom<PolicyInfo>>(
&self,
id: ftrace::Id,
) -> T {
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Read(T::POLICY_TYPE.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Read(
StorageInfo::PolicyInfo(policy_info),
)) = payload
{
let policy_type: PolicyType = (&policy_info).into();
if let Ok(info) = policy_info.try_into() {
return info;
}
panic!(
"Mismatching type during read. Expected {:?}, but got {:?}",
T::POLICY_TYPE,
policy_type
);
} else {
panic!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
panic!("Error reading from storage: {:?}", err);
}
}
}
/// Write a policy info object to storage. The argument `write_through` will
/// block returning until the value has been completely written to
/// persistent store, rather than any temporary in-memory caching.
pub(crate) async fn
|
(
&self,
policy_info: PolicyInfo,
id: ftrace::Id,
) -> Result<UpdateState, PolicyError> {
let policy_type = (&policy_info).into();
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Write(policy_info.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Write(result)) = payload
{
return result.map_err(|e| {
fx_log_err!("Failed to write policy: {:?}", e);
PolicyError::WriteFailure(policy_type)
});
} else {
fx_log_err!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
fx_log_err!("Error writing to storage: {:?}", err);
}
}
Err(PolicyError::WriteFailure(policy_type))
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::base::SettingType;
use crate::handler::base::{Payload as HandlerPayload, Request};
use crate::message::base::MessengerType;
use crate::message::MessageHubUtil;
use crate::policy::PolicyType;
use crate::service;
use crate::tests::message_utils::verify_payload;
#[fuchsia_async::run_until_stalled(test)]
async fn test_client_proxy_send_setting_request() {
let policy_type = PolicyType::Unknown;
let setting_request = Request::Get;
let target_setting_type = SettingType::Unknown;
let service_delegate = service::MessageHub::create_hub();
let (_, mut setting_proxy_receptor) = service_delegate
.create(MessengerType::Addressable(service::Address::Handler(
policy_type.setting_type(),
)))
.await
.expect("setting proxy messenger created");
let client_proxy = ClientProxy {
service_messenger: service_delegate
.create(MessengerType::Unbound)
.await
.expect("messenger should be created")
.0,
};
let _ = client_proxy.send_setting_request(target_setting_type, setting_request.clone());
verify_payload(
service::Payload::Setting(HandlerPayload::Request(setting_request)),
|
write_policy
|
identifier_name
|
policy_handler.rs
|
Factory};
use crate::message::base::Audience;
use crate::policy::response::{Error as PolicyError, Response};
use crate::policy::{
BoxedHandler, Context, GenerateHandlerResult, HasPolicyType, PolicyInfo, PolicyType,
Request as PolicyRequest,
};
use crate::service;
use crate::storage::{self, StorageInfo};
use anyhow::Error;
use async_trait::async_trait;
use fuchsia_syslog::fx_log_err;
use fuchsia_trace as ftrace;
use futures::future::BoxFuture;
use settings_storage::device_storage::DeviceStorage;
use settings_storage::UpdateState;
use std::convert::{TryFrom, TryInto};
/// PolicyHandlers are in charge of applying and persisting policies set by clients.
#[async_trait]
pub trait PolicyHandler {
/// Called when a policy client makes a request on the policy API this handler controls.
async fn handle_policy_request(&mut self, request: PolicyRequest) -> Response;
/// Called when a setting request is intercepted for the setting this policy handler supervises.
///
/// If there are no policies or the request does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to consume the request and respond to the client directly, it should
/// return [`RequestTransform::Result`].
///
/// If this handler wants to modify the request, then let the setting handler handle it,
/// [`RequestTransform::Request`] should be returned, with the modified request.
///
/// [`RequestTransform::Result`]: enum.RequestTransform.html
/// [`RequestTransform::Request`]: enum.RequestTransform.html
async fn handle_setting_request(&mut self, request: Request) -> Option<RequestTransform>;
/// Called when a setting response is intercepted from the setting this policy handler
/// supervises.
///
/// If there are no policies or the response does not need to be modified, `None` should be
/// returned.
///
/// If this handler wants to modify the response and still let the original audience handle it,
/// [`Response`] should be returned, containing the modified response.
///
/// [`Response`]: ResponseTransform::Response
async fn handle_setting_response(
&mut self,
response: SettingResponse,
) -> Option<ResponseTransform>;
}
/// `RequestTransform` is returned by a [`PolicyHandler`] in response to a setting request that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum RequestTransform {
/// A new, modified request that should be forwarded to the setting handler for processing.
Request(Request),
/// A result to return directly to the settings client.
Result(SettingHandlerResult),
}
/// `ResponseTransform` is returned by a [`PolicyHandler`] in response to a setting response that a
/// [`PolicyProxy`] intercepted. The presence of this value indicates that the policy handler has
/// decided to take action in order to apply policies.
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`PolicyProxy`]: ../policy_proxy/struct.PolicyProxy.html
///
#[derive(Clone, Debug, PartialEq)]
pub enum ResponseTransform {
/// A new, modified response that should be forwarded.
Response(SettingResponse),
}
/// Trait used to create policy handlers.
#[async_trait]
pub trait Create: Sized {
async fn create(handler: ClientProxy) -> Result<Self, Error>;
}
/// Creates a [`PolicyHandler`] from the given [`Context`].
///
/// [`PolicyHandler`]: trait.PolicyHandler.html
/// [`Context`]: ../base/struct.Context.html
pub(crate) fn create_handler<C, T: StorageFactory<Storage = DeviceStorage> + 'static>(
context: Context<T>,
) -> BoxFuture<'static, GenerateHandlerResult>
where
C: Create + PolicyHandler + Send + Sync + 'static,
{
Box::pin(async move {
let _ = &context;
let proxy = ClientProxy::new(context.service_messenger);
C::create(proxy).await.map(|handler| Box::new(handler) as BoxedHandler)
})
}
/// `ClientProxy` provides common functionality, like messaging and persistence to policy handlers.
#[derive(Clone)]
pub struct ClientProxy {
service_messenger: service::message::Messenger,
}
|
impl ClientProxy {
/// Sends a setting request to the underlying setting proxy this policy handler controls.
pub(crate) fn send_setting_request(
&self,
setting_type: SettingType,
request: Request,
) -> service::message::Receptor {
self.service_messenger
.message(
HandlerPayload::Request(request).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send()
}
/// Requests the setting handler to rebroadcast a settings changed event to its listeners.
pub(crate) fn request_rebroadcast(&self, setting_type: SettingType) {
// Ignore the receptor result.
let _ = self
.service_messenger
.message(
HandlerPayload::Request(Request::Rebroadcast).into(),
Audience::Address(service::Address::Handler(setting_type)),
)
.send();
}
}
impl ClientProxy {
pub(crate) fn new(service_messenger: service::message::Messenger) -> Self {
Self { service_messenger }
}
/// The type `T` is any type that has a [`PolicyType`] associated with it and that can be
/// converted into a [`PolicyInfo`]. This is usually a variant of the `PolicyInfo` enum.
pub(crate) async fn read_policy<T: HasPolicyType + TryFrom<PolicyInfo>>(
&self,
id: ftrace::Id,
) -> T {
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Read(T::POLICY_TYPE.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Read(
StorageInfo::PolicyInfo(policy_info),
)) = payload
{
let policy_type: PolicyType = (&policy_info).into();
if let Ok(info) = policy_info.try_into() {
return info;
}
panic!(
"Mismatching type during read. Expected {:?}, but got {:?}",
T::POLICY_TYPE,
policy_type
);
} else {
panic!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
panic!("Error reading from storage: {:?}", err);
}
}
}
/// Write a policy info object to storage. The argument `write_through` will
/// block returning until the value has been completely written to
/// persistent store, rather than any temporary in-memory caching.
pub(crate) async fn write_policy(
&self,
policy_info: PolicyInfo,
id: ftrace::Id,
) -> Result<UpdateState, PolicyError> {
let policy_type = (&policy_info).into();
let mut receptor = self
.service_messenger
.message(
storage::Payload::Request(storage::StorageRequest::Write(policy_info.into(), id))
.into(),
Audience::Address(service::Address::Storage),
)
.send();
match receptor.next_of::<storage::Payload>().await {
Ok((payload, _)) => {
if let storage::Payload::Response(storage::StorageResponse::Write(result)) = payload
{
return result.map_err(|e| {
fx_log_err!("Failed to write policy: {:?}", e);
PolicyError::WriteFailure(policy_type)
});
} else {
fx_log_err!("Incorrect response received from storage: {:?}", payload);
}
}
Err(err) => {
fx_log_err!("Error writing to storage: {:?}", err);
}
}
Err(PolicyError::WriteFailure(policy_type))
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::base::SettingType;
use crate::handler::base::{Payload as HandlerPayload, Request};
use crate::message::base::MessengerType;
use crate::message::MessageHubUtil;
use crate::policy::PolicyType;
use crate::service;
use crate::tests::message_utils::verify_payload;
#[fuchsia_async::run_until_stalled(test)]
async fn test_client_proxy_send_setting_request() {
let policy_type = PolicyType::Unknown;
let setting_request = Request::Get;
let target_setting_type = SettingType::Unknown;
let service_delegate = service::MessageHub::create_hub();
let (_, mut setting_proxy_receptor) = service_delegate
.create(MessengerType::Addressable(service::Address::Handler(
policy_type.setting_type(),
)))
.await
.expect("setting proxy messenger created");
let client_proxy = ClientProxy {
service_messenger: service_delegate
.create(MessengerType::Unbound)
.await
.expect("messenger should be created")
.0,
};
let _ = client_proxy.send_setting_request(target_setting_type, setting_request.clone());
verify_payload(
service::Payload::Setting(HandlerPayload::Request(setting_request)),
|
random_line_split
|
|
projectsvmglove.py
|
elif w == 'goood':
return 'very good'
elif w == 'tera':
return 'teraa'
elif w == 'cnfsn':
return 'confusion'
elif w == 'ka':
return 'kaa'
elif w == 'rkhi':
return 'rakhi'
elif w == 'thts':
return 'thats'
elif w == 'cald':
return 'called'
elif w == 'tabhe':
return 'tabhi'
elif w == 'pta':
return 'pata'
elif w == 'b':
return 'bhi'
elif w == 'nai':
return 'nahi'
elif w == 'f':
return 'of'
elif w == 'd':
return 'the'
else:
return w
'''
Translate word .
'''
def translate(word):
return translator.translate(word,src='hi' , dest='en').text
'''
Self defined contractions
'''
def load_dict_contractions():
return {
"ain't":"is not",
"amn't":"am not",
"aren't":"are not",
"can't":"cannot",
"'cause":"because",
"couldn't":"could not",
"couldn't've":"could not have",
"could've":"could have",
"daren't":"dare not",
"daresn't":"dare not",
"dasn't":"dare not",
"didn't":"did not",
"doesn't":"does not",
"don't":"do not",
"e'er":"ever",
"em":"them",
"everyone's":"everyone is",
"finna":"fixing to",
"gimme":"give me",
"gonna":"going to",
"gon't":"go not",
"gotta":"got to",
"hadn't":"had not",
"hasn't":"has not",
"haven't":"have not",
"he'd":"he would",
"he'll":"he will",
"he's":"he is",
"he've":"he have",
"how'd":"how would",
"how'll":"how will",
"how're":"how are",
"how's":"how is",
"I'd":"I would",
"I'll":"I will",
"i'll":"I will",
"I'm":"I am",
"I'm'a":"I am about to",
"I'm'o":"I am going to",
"isn't":"is not",
"it'd":"it would",
"it'll":"it will",
"it's":"it is",
"I've":"I have",
"kinda":"kind of",
"let's":"let us",
"mayn't":"may not",
"may've":"may have",
"mightn't":"might not",
"might've":"might have",
"mustn't":"must not",
"mustn't've":"must not have",
"must've":"must have",
"needn't":"need not",
"ne'er":"never",
"o'":"of",
"o'er":"over",
"ol'":"old",
"oughtn't":"ought not",
"shalln't":"shall not",
"shan't":"shall not",
"she'd":"she would",
"she'll":"she will",
"she's":"she is",
"shouldn't":"should not",
"shouldn't've":"should not have",
"should've":"should have",
"somebody's":"somebody is",
"someone's":"someone is",
"something's":"something is",
"that'd":"that would",
"that'll":"that will",
"that're":"that are",
"that's":"that is",
"there'd":"there would",
"there'll":"there will",
"there're":"there are",
"there's":"there is",
"these're":"these are",
"they'd":"they would",
"they'll":"they will",
"they're":"they are",
"they've":"they have",
"this's":"this is",
"those're":"those are",
"'tis":"it is",
"'twas":"it was",
"wanna":"want to",
"wasn't":"was not",
"we'd":"we would",
"we'd've":"we would have",
"we'll":"we will",
"we're":"we are",
"weren't":"were not",
"we've":"we have",
"what'd":"what did",
"what'll":"what will",
"what're":"what are",
"what's":"what is",
"what've":"what have",
"when's":"when is",
"where'd":"where did",
"where're":"where are",
"where's":"where is",
"where've":"where have",
"which's":"which is",
"who'd":"who would",
"who'd've":"who would have",
"who'll":"who will",
"who're":"who are",
"who's":"who is",
"who've":"who have",
"why'd":"why did",
"why're":"why are",
"why's":"why is",
"won't":"will not",
"wouldn't":"would not",
"would've":"would have",
"y'all":"you all",
"you'd":"you would",
"you'll":"you will",
"you're":"you are",
"you've":"you have",
"Whatcha":"What are you",
"whatcha":"What are you",
"luv":"love",
"sux":"sucks"
}
'''
Handling short forms and contractions in the sentences
'''
long_form_dict = load_dict_contractions()
def expand_sent(sentence):
final_sent =""
res = " ".join(long_form_dict.get(ele, ele) for ele in sentence.split())
for word in res.split():
final_sent += (handle_short_forms(word))+ " "
return final_sent
data.shape
l=[]
for sent in data.sentence_eng:
l.append(expand_sent(sent))
print(l[2:5])
data['sentence_eng']=l
'''
Emoticon processing
'''
emoji_list = pd.read_csv('emoji.csv',sep=",")
def find_emoji(sentence):
positive = 0
negative = 0
neutral = 0
sentiment =[]
for word in sentence.split():
if not emoji_list[emoji_list['Emoji']==word].empty:
positive += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Positive']
negative += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Negative']
neutral += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Neutral']
return positive,negative,neutral
'''
Slang word processing
'''
slang_list = pd.read_csv('Hinglish_Profanity_List.csv',sep=",",header=None)
slang_list.columns=['hindi_word','meaning','rating']
def find_slang(sentence):
slang_exists = False
total_rating = 0
for word in sentence.split():
if not slang_list[slang_list['hindi_word']==word].empty:
try:
total_rating += slang_list.iloc[slang_list.index[slang_list['hindi_word'] == word].tolist()[0]]['rating']
except:
total_rating +=0
slang_exists = True
try:
total_rating = int(total_rating)
except:
total_rating = 0
return slang_exists,total_rating
data['slang_exists']= data['sentence_mixed'].apply(find_slang)
data[['slang_existance', 'slang_rating']] = pd.DataFrame(data['slang_exists'].tolist(), index=data.index)
data =data.drop(columns=['slang_exists'])
data['emoji_sentiment'] = data['sentence_mixed'].apply(find_emoji)
data[['positive_emoji', 'negative_emoji','neutral_emoji']] = pd.DataFrame(data['emoji_sentiment'].tolist(), index=data.index)
data =data.drop(columns=['emoji_sentiment'])
'''
Calculating the sentiment polarity score
'''
sid = SentimentIntensityAnalyzer()
neg_score=[]
pos_score=[]
neu_score=[]
for sent in data.sentence_eng:
ss = sid.polarity_scores(sent)
neg_score.append(ss['neg'])
pos_score.append(ss['pos'])
neu_score.append(ss['neu'])
data['neg_score']= neg_score
data['pos_score']= pos_score
data['neu_score']= neu_score
data
|
if w == 'h':
return 'hai'
elif w == 'n':
return 'na'
elif w == 'da':
return 'the'
elif w == 'wid':
return 'with'
elif w == 'pr':
return 'par'
elif w == 'mattt':
return 'mat'
elif w == 'vo':
return 'woh'
elif w == 'ki':
return 'kee'
elif w == 'ap':
return 'aap'
elif w == 'bs':
return 'bas'
|
identifier_body
|
|
projectsvmglove.py
|
":"she will",
"she's":"she is",
"shouldn't":"should not",
"shouldn't've":"should not have",
"should've":"should have",
"somebody's":"somebody is",
"someone's":"someone is",
"something's":"something is",
"that'd":"that would",
"that'll":"that will",
"that're":"that are",
"that's":"that is",
"there'd":"there would",
"there'll":"there will",
"there're":"there are",
"there's":"there is",
"these're":"these are",
"they'd":"they would",
"they'll":"they will",
"they're":"they are",
"they've":"they have",
"this's":"this is",
"those're":"those are",
"'tis":"it is",
"'twas":"it was",
"wanna":"want to",
"wasn't":"was not",
"we'd":"we would",
"we'd've":"we would have",
"we'll":"we will",
"we're":"we are",
"weren't":"were not",
"we've":"we have",
"what'd":"what did",
"what'll":"what will",
"what're":"what are",
"what's":"what is",
"what've":"what have",
"when's":"when is",
"where'd":"where did",
"where're":"where are",
"where's":"where is",
"where've":"where have",
"which's":"which is",
"who'd":"who would",
"who'd've":"who would have",
"who'll":"who will",
"who're":"who are",
"who's":"who is",
"who've":"who have",
"why'd":"why did",
"why're":"why are",
"why's":"why is",
"won't":"will not",
"wouldn't":"would not",
"would've":"would have",
"y'all":"you all",
"you'd":"you would",
"you'll":"you will",
"you're":"you are",
"you've":"you have",
"Whatcha":"What are you",
"whatcha":"What are you",
"luv":"love",
"sux":"sucks"
}
'''
Handling short forms and contractions in the sentences
'''
long_form_dict = load_dict_contractions()
def expand_sent(sentence):
final_sent =""
res = " ".join(long_form_dict.get(ele, ele) for ele in sentence.split())
for word in res.split():
final_sent += (handle_short_forms(word))+ " "
return final_sent
data.shape
l=[]
for sent in data.sentence_eng:
l.append(expand_sent(sent))
print(l[2:5])
data['sentence_eng']=l
'''
Emoticon processing
'''
emoji_list = pd.read_csv('emoji.csv',sep=",")
def find_emoji(sentence):
positive = 0
negative = 0
neutral = 0
sentiment =[]
for word in sentence.split():
if not emoji_list[emoji_list['Emoji']==word].empty:
positive += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Positive']
negative += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Negative']
neutral += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Neutral']
return positive,negative,neutral
'''
Slang word processing
'''
slang_list = pd.read_csv('Hinglish_Profanity_List.csv',sep=",",header=None)
slang_list.columns=['hindi_word','meaning','rating']
def find_slang(sentence):
slang_exists = False
total_rating = 0
for word in sentence.split():
if not slang_list[slang_list['hindi_word']==word].empty:
try:
total_rating += slang_list.iloc[slang_list.index[slang_list['hindi_word'] == word].tolist()[0]]['rating']
except:
total_rating +=0
slang_exists = True
try:
total_rating = int(total_rating)
except:
total_rating = 0
return slang_exists,total_rating
data['slang_exists']= data['sentence_mixed'].apply(find_slang)
data[['slang_existance', 'slang_rating']] = pd.DataFrame(data['slang_exists'].tolist(), index=data.index)
data =data.drop(columns=['slang_exists'])
data['emoji_sentiment'] = data['sentence_mixed'].apply(find_emoji)
data[['positive_emoji', 'negative_emoji','neutral_emoji']] = pd.DataFrame(data['emoji_sentiment'].tolist(), index=data.index)
data =data.drop(columns=['emoji_sentiment'])
'''
Calculating the sentiment polarity score
'''
sid = SentimentIntensityAnalyzer()
neg_score=[]
pos_score=[]
neu_score=[]
for sent in data.sentence_eng:
ss = sid.polarity_scores(sent)
neg_score.append(ss['neg'])
pos_score.append(ss['pos'])
neu_score.append(ss['neu'])
data['neg_score']= neg_score
data['pos_score']= pos_score
data['neu_score']= neu_score
data.tail(10)
# removing unwanted patterns from the data
import re
import nltk
nltk.download('stopwords')
from nltk.corpus import stopwords
from nltk.stem.porter import PorterStemmer
#Splitting the data
train = data[:5000]
test = data[5000:]
'''
fetching the sentence/comments from the test dataset and cleaning it
'''
train_corpus = []
for i in range(0, 5000):
review = re.sub('[^a-zA-Z]', ' ', train['sentence_eng'][i])
review = review.lower()
review = review.split()
ps = PorterStemmer()
# stemming
review = [ps.stem(word) for word in review if not word in set(stopwords.words('english'))]
# joining them back with space
review = ' '.join(review)
train_corpus.append(review)
'''
fetching the sentence/comments from the test dataset and cleaning it
'''
test_corpus = []
for i in range(5000, 6137):
review = re.sub('[^a-zA-Z]', ' ', test['sentence_eng'][i])
review = review.lower()
review = review.split()
ps = PorterStemmer()
# stemming
review = [ps.stem(word) for word in review if not word in set(stopwords.words('english'))]
# joining them back with space
review = ' '.join(review)
test_corpus.append(review)
#Conversion of snetiment into numerical sentiment
train =train.replace("positive",1)
train =train.replace("negative",2)
train = train.replace("neutral",0)
train.head()
y = train.iloc[:, 2]
print(y)
from nltk.tokenize import WordPunctTokenizer
tokenizer = WordPunctTokenizer()
data_tok = [tokenizer.tokenize(d.lower()) for d in train_corpus]
import gensim.downloader as api
model = api.load('glove-twitter-50')
import numpy as np
def get_phrase_embedding(phrase):
"""
Convert phrase to a vector by aggregating it's word embeddings. Just take an
average of vectors for all tokens in the phrase with some weights.
"""
vector = np.zeros([model.vector_size], dtype='float32')
# 1. lowercase phrase
phrase = phrase.lower()
# 2. tokenize phrase
phrase_tokens = tokenizer.tokenize(phrase)
# 3. average word vectors for all words in tokenized phrase, skip words that are not in model's vocabulary
divisor = 0
for word in phrase_tokens:
if word in model.vocab:
divisor += 1
vector = vector + model.get_vector(word)
if divisor != 0: vector /= divisor
return vector
'''
Embeddings for Training Corpus
'''
vector_matrix_x_train = list(map(get_phrase_embedding, train_corpus))
'''
Embeddings for Training Corpus
'''
vector_matrix_x_test = list(map(get_phrase_embedding, test_corpus))
numerical_features_t = train[['neg_score','pos_score', 'neu_score', 'positive_emoji', 'negative_emoji', 'neutral_emoji', 'slang_rating']]
print(numerical_features_t.values.tolist())
'''
Stacking numerical features along with textual features
'''
combinedFeatures = np.hstack([numerical_features_t, np.array(vector_matrix_x_train)])
numerical_features_test = test[['neg_score','pos_score', 'neu_score', 'positive_emoji', 'negative_emoji', 'neutral_emoji', 'slang_rating']]
print(numerical_features_test.values.tolist())
combinedFeatures_test = np.hstack([numerical_features_test, np.array(vector_matrix_x_test)])
from sklearn.model_selection import train_test_split
'''
Train test validation split
'''
x_train, x_valid, y_train, y_valid = train_test_split(combinedFeatures, y, test_size = 0.25, random_state = 42)
print(x_train.shape)
|
print(x_valid.shape)
print(y_train.shape)
print(y_valid.shape)
print(combinedFeatures_test.shape)
|
random_line_split
|
|
projectsvmglove.py
|
elif w == 'goood':
return 'very good'
elif w == 'tera':
return 'teraa'
elif w == 'cnfsn':
return 'confusion'
elif w == 'ka':
return 'kaa'
elif w == 'rkhi':
return 'rakhi'
elif w == 'thts':
return 'thats'
elif w == 'cald':
return 'called'
elif w == 'tabhe':
return 'tabhi'
elif w == 'pta':
return 'pata'
elif w == 'b':
return 'bhi'
elif w == 'nai':
return 'nahi'
elif w == 'f':
return 'of'
elif w == 'd':
return 'the'
else:
return w
'''
Translate word .
'''
def translate(word):
return translator.translate(word,src='hi' , dest='en').text
'''
Self defined contractions
'''
def load_dict_contractions():
return {
"ain't":"is not",
"amn't":"am not",
"aren't":"are not",
"can't":"cannot",
"'cause":"because",
"couldn't":"could not",
"couldn't've":"could not have",
"could've":"could have",
"daren't":"dare not",
"daresn't":"dare not",
"dasn't":"dare not",
"didn't":"did not",
"doesn't":"does not",
"don't":"do not",
"e'er":"ever",
"em":"them",
"everyone's":"everyone is",
"finna":"fixing to",
"gimme":"give me",
"gonna":"going to",
"gon't":"go not",
"gotta":"got to",
"hadn't":"had not",
"hasn't":"has not",
"haven't":"have not",
"he'd":"he would",
"he'll":"he will",
"he's":"he is",
"he've":"he have",
"how'd":"how would",
"how'll":"how will",
"how're":"how are",
"how's":"how is",
"I'd":"I would",
"I'll":"I will",
"i'll":"I will",
"I'm":"I am",
"I'm'a":"I am about to",
"I'm'o":"I am going to",
"isn't":"is not",
"it'd":"it would",
"it'll":"it will",
"it's":"it is",
"I've":"I have",
"kinda":"kind of",
"let's":"let us",
"mayn't":"may not",
"may've":"may have",
"mightn't":"might not",
"might've":"might have",
"mustn't":"must not",
"mustn't've":"must not have",
"must've":"must have",
"needn't":"need not",
"ne'er":"never",
"o'":"of",
"o'er":"over",
"ol'":"old",
"oughtn't":"ought not",
"shalln't":"shall not",
"shan't":"shall not",
"she'd":"she would",
"she'll":"she will",
"she's":"she is",
"shouldn't":"should not",
"shouldn't've":"should not have",
"should've":"should have",
"somebody's":"somebody is",
"someone's":"someone is",
"something's":"something is",
"that'd":"that would",
"that'll":"that will",
"that're":"that are",
"that's":"that is",
"there'd":"there would",
"there'll":"there will",
"there're":"there are",
"there's":"there is",
"these're":"these are",
"they'd":"they would",
"they'll":"they will",
"they're":"they are",
"they've":"they have",
"this's":"this is",
"those're":"those are",
"'tis":"it is",
"'twas":"it was",
"wanna":"want to",
"wasn't":"was not",
"we'd":"we would",
"we'd've":"we would have",
"we'll":"we will",
"we're":"we are",
"weren't":"were not",
"we've":"we have",
"what'd":"what did",
"what'll":"what will",
"what're":"what are",
"what's":"what is",
"what've":"what have",
"when's":"when is",
"where'd":"where did",
"where're":"where are",
"where's":"where is",
"where've":"where have",
"which's":"which is",
"who'd":"who would",
"who'd've":"who would have",
"who'll":"who will",
"who're":"who are",
"who's":"who is",
"who've":"who have",
"why'd":"why did",
"why're":"why are",
"why's":"why is",
"won't":"will not",
"wouldn't":"would not",
"would've":"would have",
"y'all":"you all",
"you'd":"you would",
"you'll":"you will",
"you're":"you are",
"you've":"you have",
"Whatcha":"What are you",
"whatcha":"What are you",
"luv":"love",
"sux":"sucks"
}
'''
Handling short forms and contractions in the sentences
'''
long_form_dict = load_dict_contractions()
def expand_sent(sentence):
final_sent =""
res = " ".join(long_form_dict.get(ele, ele) for ele in sentence.split())
for word in res.split():
final_sent += (handle_short_forms(word))+ " "
return final_sent
data.shape
l=[]
for sent in data.sentence_eng:
l.append(expand_sent(sent))
print(l[2:5])
data['sentence_eng']=l
'''
Emoticon processing
'''
emoji_list = pd.read_csv('emoji.csv',sep=",")
def find_emoji(sentence):
positive = 0
negative = 0
neutral = 0
sentiment =[]
for word in sentence.split():
if not emoji_list[emoji_list['Emoji']==word].empty:
positive += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Positive']
negative += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Negative']
neutral += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Neutral']
return positive,negative,neutral
'''
Slang word processing
'''
slang_list = pd.read_csv('Hinglish_Profanity_List.csv',sep=",",header=None)
slang_list.columns=['hindi_word','meaning','rating']
def find_slang(sentence):
slang_exists = False
total_rating = 0
for word in sentence.split():
if not slang_list[slang_list['hindi_word']==word].empty:
try:
total_rating += slang_list.iloc[slang_list.index[slang_list['hindi_word'] == word].tolist()[0]]['rating']
except:
total_rating +=0
slang_exists = True
try:
total_rating = int(total_rating)
except:
total_rating = 0
return slang_exists,total_rating
data['slang_exists']= data['sentence_mixed'].apply(find_slang)
data[['slang_existance', 'slang_rating']] = pd.DataFrame(data['slang_exists'].tolist(), index=data.index)
data =data.drop(columns=['slang_exists'])
data['emoji_sentiment'] = data['sentence_mixed'].apply(find_emoji)
data[['positive_emoji', 'negative_emoji','neutral_emoji']] = pd.DataFrame(data['emoji_sentiment'].tolist(), index=data.index)
data =data.drop(columns=['emoji_sentiment'])
'''
Calculating the sentiment polarity score
'''
sid = SentimentIntensityAnalyzer()
neg_score=[]
pos_score=[]
neu_score=[]
for sent in data.sentence_eng:
ss = sid.polarity_scores(sent)
neg_score.append(ss['neg'])
pos_score.append(ss['pos'])
neu_score.append(ss['neu'])
data['neg_score']= neg_score
data['pos_score']= pos_score
data['neu_score']= neu_score
data.tail(10)
# removing unwanted patterns from the data
import re
import nltk
nltk.download('stopwords')
from nltk.corpus import stopwords
from nltk.stem.porter import PorterStemmer
#Splitting the data
train = data[:5000]
test = data[5000:]
'''
fetching the sentence/comments from the test dataset and cleaning it
'''
train_corpus = []
for i in range(0, 5000):
review = re.sub('[^a-zA-Z]', ' ', train['sentence_eng'][i
|
return 'bas'
|
conditional_block
|
|
projectsvmglove.py
|
ata'
elif w == 'b':
return 'bhi'
elif w == 'nai':
return 'nahi'
elif w == 'f':
return 'of'
elif w == 'd':
return 'the'
else:
return w
'''
Translate word .
'''
def translate(word):
return translator.translate(word,src='hi' , dest='en').text
'''
Self defined contractions
'''
def load_dict_contractions():
return {
"ain't":"is not",
"amn't":"am not",
"aren't":"are not",
"can't":"cannot",
"'cause":"because",
"couldn't":"could not",
"couldn't've":"could not have",
"could've":"could have",
"daren't":"dare not",
"daresn't":"dare not",
"dasn't":"dare not",
"didn't":"did not",
"doesn't":"does not",
"don't":"do not",
"e'er":"ever",
"em":"them",
"everyone's":"everyone is",
"finna":"fixing to",
"gimme":"give me",
"gonna":"going to",
"gon't":"go not",
"gotta":"got to",
"hadn't":"had not",
"hasn't":"has not",
"haven't":"have not",
"he'd":"he would",
"he'll":"he will",
"he's":"he is",
"he've":"he have",
"how'd":"how would",
"how'll":"how will",
"how're":"how are",
"how's":"how is",
"I'd":"I would",
"I'll":"I will",
"i'll":"I will",
"I'm":"I am",
"I'm'a":"I am about to",
"I'm'o":"I am going to",
"isn't":"is not",
"it'd":"it would",
"it'll":"it will",
"it's":"it is",
"I've":"I have",
"kinda":"kind of",
"let's":"let us",
"mayn't":"may not",
"may've":"may have",
"mightn't":"might not",
"might've":"might have",
"mustn't":"must not",
"mustn't've":"must not have",
"must've":"must have",
"needn't":"need not",
"ne'er":"never",
"o'":"of",
"o'er":"over",
"ol'":"old",
"oughtn't":"ought not",
"shalln't":"shall not",
"shan't":"shall not",
"she'd":"she would",
"she'll":"she will",
"she's":"she is",
"shouldn't":"should not",
"shouldn't've":"should not have",
"should've":"should have",
"somebody's":"somebody is",
"someone's":"someone is",
"something's":"something is",
"that'd":"that would",
"that'll":"that will",
"that're":"that are",
"that's":"that is",
"there'd":"there would",
"there'll":"there will",
"there're":"there are",
"there's":"there is",
"these're":"these are",
"they'd":"they would",
"they'll":"they will",
"they're":"they are",
"they've":"they have",
"this's":"this is",
"those're":"those are",
"'tis":"it is",
"'twas":"it was",
"wanna":"want to",
"wasn't":"was not",
"we'd":"we would",
"we'd've":"we would have",
"we'll":"we will",
"we're":"we are",
"weren't":"were not",
"we've":"we have",
"what'd":"what did",
"what'll":"what will",
"what're":"what are",
"what's":"what is",
"what've":"what have",
"when's":"when is",
"where'd":"where did",
"where're":"where are",
"where's":"where is",
"where've":"where have",
"which's":"which is",
"who'd":"who would",
"who'd've":"who would have",
"who'll":"who will",
"who're":"who are",
"who's":"who is",
"who've":"who have",
"why'd":"why did",
"why're":"why are",
"why's":"why is",
"won't":"will not",
"wouldn't":"would not",
"would've":"would have",
"y'all":"you all",
"you'd":"you would",
"you'll":"you will",
"you're":"you are",
"you've":"you have",
"Whatcha":"What are you",
"whatcha":"What are you",
"luv":"love",
"sux":"sucks"
}
'''
Handling short forms and contractions in the sentences
'''
long_form_dict = load_dict_contractions()
def
|
(sentence):
final_sent =""
res = " ".join(long_form_dict.get(ele, ele) for ele in sentence.split())
for word in res.split():
final_sent += (handle_short_forms(word))+ " "
return final_sent
data.shape
l=[]
for sent in data.sentence_eng:
l.append(expand_sent(sent))
print(l[2:5])
data['sentence_eng']=l
'''
Emoticon processing
'''
emoji_list = pd.read_csv('emoji.csv',sep=",")
def find_emoji(sentence):
positive = 0
negative = 0
neutral = 0
sentiment =[]
for word in sentence.split():
if not emoji_list[emoji_list['Emoji']==word].empty:
positive += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Positive']
negative += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Negative']
neutral += emoji_list.iloc[emoji_list.index[emoji_list['Emoji'] == word].tolist()[0]]['Neutral']
return positive,negative,neutral
'''
Slang word processing
'''
slang_list = pd.read_csv('Hinglish_Profanity_List.csv',sep=",",header=None)
slang_list.columns=['hindi_word','meaning','rating']
def find_slang(sentence):
slang_exists = False
total_rating = 0
for word in sentence.split():
if not slang_list[slang_list['hindi_word']==word].empty:
try:
total_rating += slang_list.iloc[slang_list.index[slang_list['hindi_word'] == word].tolist()[0]]['rating']
except:
total_rating +=0
slang_exists = True
try:
total_rating = int(total_rating)
except:
total_rating = 0
return slang_exists,total_rating
data['slang_exists']= data['sentence_mixed'].apply(find_slang)
data[['slang_existance', 'slang_rating']] = pd.DataFrame(data['slang_exists'].tolist(), index=data.index)
data =data.drop(columns=['slang_exists'])
data['emoji_sentiment'] = data['sentence_mixed'].apply(find_emoji)
data[['positive_emoji', 'negative_emoji','neutral_emoji']] = pd.DataFrame(data['emoji_sentiment'].tolist(), index=data.index)
data =data.drop(columns=['emoji_sentiment'])
'''
Calculating the sentiment polarity score
'''
sid = SentimentIntensityAnalyzer()
neg_score=[]
pos_score=[]
neu_score=[]
for sent in data.sentence_eng:
ss = sid.polarity_scores(sent)
neg_score.append(ss['neg'])
pos_score.append(ss['pos'])
neu_score.append(ss['neu'])
data['neg_score']= neg_score
data['pos_score']= pos_score
data['neu_score']= neu_score
data.tail(10)
# removing unwanted patterns from the data
import re
import nltk
nltk.download('stopwords')
from nltk.corpus import stopwords
from nltk.stem.porter import PorterStemmer
#Splitting the data
train = data[:5000]
test = data[5000:]
'''
fetching the sentence/comments from the test dataset and cleaning it
'''
train_corpus = []
for i in range(0, 5000):
review = re.sub('[^a-zA-Z]', ' ', train['sentence_eng'][i])
review = review.lower()
review = review.split()
ps = PorterStemmer()
# stemming
review = [ps.stem(word) for word in review if not word in set(stopwords.words('english'))]
# joining them back with space
review = ' '.join(review)
train_corpus.append(review)
'''
fetching the sentence/comments from the test dataset and cleaning it
'''
test_corpus = []
for i in range(5000, 6137):
review = re.sub('[^a-zA-Z]', ' ', test['sentence
|
expand_sent
|
identifier_name
|
main.py
|
msg': 'Requisição enviada para aprovação!'}
else:
json_ret = {'status': 500, 'msg': msg}
else:
json_ret = {'status': 500, 'msg': 'Json inválido!'}
return json_ret
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def delete_card_request_handler(event, context):
"""
DELETE - Esse código será chamado através de um DELETE para a API que será criada no arquivo serverless.
Faz a requisição para apagar uma solicitação de cartão.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi excluído ou não.
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
del_response, msg = s3.delete_s3_obj('', path['id'])
if del_response['ResponseMetadata']['HTTPStatusCode'] == 204:
return {'status': 200, 'msg': 'Requisição deletada!'}
else:
return {'status': 404, 'msg': 'Requisição não encontrado'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def process_card_request_handler(event, context):
"""
Realiza o processamento de um pedido de novo cartão.
Esse código é ativado quando um novo arquivo de pedido é salvo no S3
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: None.
"""
try:
for obj in event['Records']:
# Informações do arquivo que foi inserido no S3
bucket_name = obj['s3']['bucket']['name']
obj = obj['s3']['object']['key']
s3 = S3(bucket_name)
s3.create_s3_instance()
obj_json, msg = s3.get_s3_obj('', obj)
# faz os Score do cliente
score = randint(1, 999)
obj_json['credit'] = score_to_text(score, obj_json['income'])
resp, msg = s3.put_s3_obj('', obj_json, obj_json['cpf'])
except Exception as err:
print({'status': 500, 'msg': 'Erro interno ao processar a requisição', "error": f'{err}'})
def get_new_card_request_response_handler(event, context):
"""
GET - Esse código será chamado através de um GET para a API que será criada no arquivo serverless.
Recebe o id do solicitante via parametro de URL e retorna as informações referente ao solicitante e seu crédito
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: informações do solicitante
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
print(event)
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
request_json, msg = s3.get_s3_obj('', path['id'])
if request_json:
request_json['status'] = 200
return request_json
else:
return {'status': 404, 'msg': 'Requisição não encontrada'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
if __name__ == '__main__':
ex_event = {'body':
{
'name': 'Leonardo Roberto Gazziro',
'phone': '99999999999',
'age': 26,
'cpf': '999.999.999-99',
'income': 3000
},
'method': 'POST', 'principalId': '',
'stage': 'dev',
'cognitoPoolClaims': {'sub': ''},
'enhancedAuthContext': {},
'headers': {'Accept': '*/*', 'Accept-Encoding': 'gzip, deflate, br',
'Cache-Control': 'no-cache', 'CloudFront-Forwarded-Proto': 'https',
'CloudFront-Is-Desktop-Viewer': 'true',
'CloudFront-Is-Mobile-Viewer': 'false',
'CloudFront-Is-SmartTV-Viewer': 'false',
'CloudFront-Is-Tablet-Viewer': 'false',
'CloudFront-Viewer-Country': 'BR',
'Content-Type': 'application/json',
'Host': '1nhucniq8b.execute-api.us-east-1.amazonaws.com',
'Postman-Token': '0d40ddc9-494c-438c-9e34-30f9f44ea018',
'User-Agent': 'PostmanRuntime/7.26.3',
'Via': '1.1 3fff6e22f8d6795a61bfdca17d362ca5.cloudfront.net (CloudFront)',
'X-Amz-Cf-Id': 'OKNx6jkzKLQ3nbtD0t4JTNynGlc2TZDSemAepsPC-8Kv0ZV1f6Tz7w==',
'X-Amzn-Trace-Id': 'Root=1-5f4bf435-7d8aa2621cebaaabfa719a0c',
'X-Forwarded-For': '138.204.24.213, 64.252.179.69',
'X-Forwarded-Port': '443', 'X-Forwarded-Proto': 'https'},
'query': {},
'path': {},
'identity': {'cognitoIdentityPoolId': '', 'accountId': '', 'cognitoIdentityId': '', 'caller': '',
'sourceIp': '138.204.24.213', 'principalOrgId': '', 'accessKey': '',
'cognitoAuthenticationType': '', 'cognitoAuthenticationProvider': '', 'userArn': '',
'userAgent': 'PostmanRuntime/7.26.3', 'user': ''},
'stageVariables': {},
'requestPath': ''}
ex_event_s3 = {'Records': [{'eventVersion': '2.1', 'eventSource': 'aws:s3', 'awsRegion': 'us-east-1',
'eventTime': '2020-09-06T18:11:57.722Z', 'eventName': 'ObjectCreated:Put',
'userIdentity': {'principalId': 'A1GN6EPM0JKA8K'},
'requestParameters': {'sourceIPAddress': '177.42.49.149'},
'responseElements': {'x-amz-request-id': '7857E87B7E4BA60D',
'x-amz-id-2': 'lmNysI3mKLmMQoOzCtPnjT8usl2fMUYbIyipfE59v3oWuyu44XxI/L2tXxRPkqjC6uUNu3rGB/eekMpWqOj6RceGfrLLLTeg'},
's3': {'s3SchemaVersion': '1.0',
'configurationId': 'cc7a7b1c-d354-413c-94e0-c4c2e3cf153a',
'bucket': {'name': 'cards-requests-leogazziro',
'ownerIdentity': {'principalId': 'A1GN6EPM0JKA8K'},
'arn': 'arn:aws:s3:::novo-produto'},
'object': {'key': '999.999.999-99.json', 'size': 1172,
'eTag': 'f64c165d9209eb645b660f04b27dc8d2',
'sequencer': '005F552670F6AF15F4'}}}]}
ex_event_att_price = {'body': {},
'method': 'GET',
|
'principalId': '',
'stage': 'dev',
'cognitoPoolClaims': {'sub': ''},
'enhancedAuthContext': {},
|
random_line_split
|
|
main.py
|
'msg': 'Erro interno ao processar a requisição'}
def request_new_card_handler(event, context):
"""
POST - Esse código se
|
print(msg)
json_ret = {'status': 200, 'msg': 'Requisição enviada para aprovação!'}
else:
json_ret = {'status': 500, 'msg': msg}
else:
json_ret = {'status': 500, 'msg': 'Json inválido!'}
return json_ret
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def delete_card_request_handler(event, context):
"""
DELETE - Esse código será chamado através de um DELETE para a API que será criada no arquivo serverless.
Faz a requisição para apagar uma solicitação de cartão.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi excluído ou não.
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
del_response, msg = s3.delete_s3_obj('', path['id'])
if del_response['ResponseMetadata']['HTTPStatusCode'] == 204:
return {'status': 200, 'msg': 'Requisição deletada!'}
else:
return {'status': 404, 'msg': 'Requisição não encontrado'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def process_card_request_handler(event, context):
"""
Realiza o processamento de um pedido de novo cartão.
Esse código é ativado quando um novo arquivo de pedido é salvo no S3
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: None.
"""
try:
for obj in event['Records']:
# Informações do arquivo que foi inserido no S3
bucket_name = obj['s3']['bucket']['name']
obj = obj['s3']['object']['key']
s3 = S3(bucket_name)
s3.create_s3_instance()
obj_json, msg = s3.get_s3_obj('', obj)
# faz os Score do cliente
score = randint(1, 999)
obj_json['credit'] = score_to_text(score, obj_json['income'])
resp, msg = s3.put_s3_obj('', obj_json, obj_json['cpf'])
except Exception as err:
print({'status': 500, 'msg': 'Erro interno ao processar a requisição', "error": f'{err}'})
def get_new_card_request_response_handler(event, context):
"""
GET - Esse código será chamado através de um GET para a API que será criada no arquivo serverless.
Recebe o id do solicitante via parametro de URL e retorna as informações referente ao solicitante e seu crédito
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: informações do solicitante
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
print(event)
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
request_json, msg = s3.get_s3_obj('', path['id'])
if request_json:
request_json['status'] = 200
return request_json
else:
return {'status': 404, 'msg': 'Requisição não encontrada'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
if __name__ == '__main__':
ex_event = {'body':
{
'name': 'Leonardo Roberto Gazziro',
'phone': '99999999999',
'age': 26,
'cpf': '999.999.999-99',
'income': 3000
},
'method': 'POST', 'principalId': '',
'stage': 'dev',
'cognitoPoolClaims': {'sub': ''},
'enhancedAuthContext': {},
'headers': {'Accept': '*/*', 'Accept-Encoding': 'gzip, deflate, br',
'Cache-Control': 'no-cache', 'CloudFront-Forwarded-Proto': 'https',
'CloudFront-Is-Desktop-Viewer': 'true',
'CloudFront-Is-Mobile-Viewer': 'false',
'CloudFront-Is-SmartTV-Viewer': 'false',
'CloudFront-Is-Tablet-Viewer': 'false',
'CloudFront-Viewer-Country': 'BR',
'Content-Type': 'application/json',
'Host': '1nhucniq8b.execute-api.us-east-1.amazonaws.com',
'Postman-Token': '0d40ddc9-494c-438c-9e34-30f9f44ea018',
'User-Agent': 'PostmanRuntime/7.26.3',
'Via': '1.1 3fff6e22f8d6795a61bfdca17d362ca5.cloudfront.net (CloudFront)',
'X-Amz-Cf-Id': 'OKNx6jkzKLQ3nbtD0t4JTNynGlc2TZDSemAepsPC-8Kv0ZV1f6Tz7w==',
'X-Amzn-Trace-Id': 'Root=1-5f4bf435-7d8aa2621cebaaabfa719a0c',
'X-Forwarded-For': '138.204.24.213, 64.252.179.69',
'X-Forwarded-Port': '443', 'X-Forwarded-Proto': 'https'},
'query': {},
'path': {},
'identity': {'cognitoIdentityPoolId': '', 'accountId': '', 'cognitoIdentityId': '', 'caller': '',
'sourceIp': '138.204.24.213', 'principalOrgId': '', 'accessKey': '',
'cognitoAuthenticationType': '', 'cognitoAuthenticationProvider': '', 'userArn': '',
'userAgent': 'PostmanRuntime/7.26.3', 'user': ''},
'stageVariables': {},
'requestPath': ''}
ex_event_s3 = {'Records': [{'eventVersion': '2.1', 'eventSource': 'aws:s3', 'awsRegion': 'us-east-1',
'eventTime': '2020-09-06T18:11:57.722Z', 'eventName': 'ObjectCreated:Put',
'userIdentity': {'principalId': 'A1GN6EPM0JKA8K'},
'requestParameters': {'sourceIPAddress': '177.42.49.149'},
'responseElements': {'x-amz-request-id': '7857E87B7E4BA60D',
'x-amz-id-2': 'lmNysI3mKLmMQoOzCtPnjT8usl2fMUYbIyipfE59v3oWuyu44XxI/L2tXxRPkqjC6uUNu3rGB/eekMpWq
|
rá chamado através de um POST para a API que será criada no arquivo serverless.
Faz a requisição de um novo cartão, o Score do candidato será avaliado.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi aprovado ou não.
"""
try:
body = event.get('body', {})
s3_bucket = getenv('S3_BUCKET', '')
if body:
# Verifica se o JSON é valido
json_valido, msg = valid_new_card_request(body)
if json_valido:
body['id'] = str(uuid.uuid4())
s3 = S3(s3_bucket)
s3.create_s3_instance()
# insere o JSON no S3
resp, msg = s3.put_s3_obj('', body, body['cpf'])
print(resp)
|
identifier_body
|
main.py
|
r:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def request_new_card_handler(event, context):
"""
POST - Esse código será chamado através de um POST para a API que será criada no arquivo serverless.
Faz a requisição de um novo cartão, o Score do candidato será avaliado.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi aprovado ou não.
"""
try:
body = event.get('body', {})
s3_bucket = getenv('S3_BUCKET', '')
if body:
# Verifica se o JSON é valido
json_valido, msg = valid_new_card_request(body)
if json_valido:
body['id'] = str(uuid.uuid4())
s3 = S3(s3_bucket)
s3.create_s3_instance()
# insere o JSON no S3
resp, msg = s3.put_s3_obj('', body, body['cpf'])
print(resp)
print(msg)
json_ret = {'status': 200, 'msg': 'Requisição enviada para aprovação!'}
else:
json_ret = {'status': 500, 'msg': msg}
else:
json_ret = {'status': 500, 'msg': 'Json inválido!'}
return json_ret
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def delete_card_request_handler(event, context):
"""
DELETE - Esse código será chamado através de um DELETE para a API que será criada no arquivo serverless.
Faz a requisição para apagar uma solicitação de cartão.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi excluído ou não.
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
del_response, msg = s3.delete_s3_obj('', path['id'])
if del_response['ResponseMetadata']['HTTPStatusCode'] == 204:
return {'status': 200, 'msg': 'Requisição deletada!'}
else:
return {'status': 404, 'msg': 'Requisição não encontrado'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def process_card_request_handler(event, context):
"""
Realiza o processamento de um pedido de novo cartão.
Esse código é ativado quando um novo arquivo de pedido é salvo no S3
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: None.
"""
try:
for obj in event['Records']:
# Informações do arquivo que foi inserido no S3
bucket_name = obj['s3']['bucket']['name']
obj = obj['s3']['object']['key']
s3 = S3(bucket_name)
s3.create_s3_instance()
obj_json, msg = s3.get_s3_obj('', obj)
# faz os Score do cliente
score = randint(1, 999)
obj_json['credit'] = score_to_text(score, obj_json['income'])
resp, msg = s3.put_s3_obj('', obj_json, obj_json['cpf'])
except Exception as err:
print({'status': 500, 'msg': 'Erro interno ao processar a requisição', "error": f'{err}'})
def get_new_card_request_response_handler(event, context):
"""
GET - Esse código será chamado através de um GET para a API que será criada no arquivo serverless.
Recebe o id do solicitante via parametro de URL e retorna as informações referente ao solicitante e seu crédito
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: informações do solicitante
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
print(event)
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
request_json, msg = s3.get_s3_obj('', path['id'])
if request_json:
request_json['status'] = 200
return request_json
else:
return {'status': 404, 'msg': 'Requisição não encontrada'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
if __name__ == '__main__':
ex_event = {'body':
{
'name': 'Leonardo Roberto Gazziro',
'phone': '99999999999',
'age': 26,
'cpf': '999.999.999-99',
'income': 3000
},
'method': 'POST', 'principalId': '',
'stage': 'dev',
'cognitoPoolClaims': {'sub': ''},
'enhancedAuthContext': {},
'headers': {'Accept': '*/*', 'Accept-Encoding': 'gzip, deflate, br',
'Cache-Control': 'no-cache', 'CloudFront-Forwarded-Proto': 'https',
'CloudFront-Is-Desktop-Viewer': 'true',
'CloudFront-Is-Mobile-Viewer': 'false',
'CloudFront-Is-SmartTV-Viewer': 'false',
'CloudFront-Is-Tablet-Viewer': 'false',
'CloudFront-Viewer-Country': 'BR',
'Content-Type': 'application/json',
'Host': '1nhucniq8b.execute-api.us-east-1.amazonaws.com',
'Postman-Token': '0d40ddc9-494c-438c-9e34-30f9f44ea018',
'User-Agent': 'PostmanRuntime/7.26.3',
'Via': '1.1 3fff6e22f8d6795a61bfdca17d362ca5.cloudfront.net (CloudFront)',
'X-Amz-Cf-Id': 'OKNx6jkzKLQ3nbtD0t4JTNynGlc2TZDSemAepsPC-8Kv0ZV1f6Tz7w==',
'X-Amzn-Trace-Id': 'Root=1-5f4bf435-7d8aa2621cebaaabfa719a0c',
'X-Forwarded-For': '138.204.24.213, 64.252.179.69',
'X-Forwarded-Port': '443', 'X-Forwarded-Proto': 'https'},
'query': {},
'path': {},
'identity': {'cognitoIdentityPoolId': '', 'accountId': '', 'cognitoIdentityId': '', 'caller': '',
'sourceIp': '138.204.24.213', 'principalOrgId': '', 'accessKey': '',
'cognitoAuthenticationType': '', 'cognitoAuthenticationProvider': '', 'userArn': '',
'userAgent': 'PostmanRuntime/7.26.3', 'user': ''},
'stageVariables': {},
'requestPath': ''}
ex_event_s3 = {'Records': [{'eventVersion': '2.1', 'eventSource': 'aws:s3', 'awsRegion': 'us-east-1',
'eventTime': '2020-09-06T18:11:57.722Z', 'eventName': 'ObjectCreated:Put',
'userIdentity': {'principalId': 'A1GN6EPM0JKA8K'},
'requestParameters': {'sourceIPAddress': '177.42.49.149'},
'responseElements': {'x-amz-request-id': '7857E87B7E4BA60D',
'x-amz-id-2': 'lmNysI3mKLmMQoOzCtPnjT8usl2fMUYbIyipfE59v3oW
|
msg': 'Lista de requisições não encontrada'}
except Exception as er
|
conditional_block
|
|
main.py
|
'msg': 'Erro interno ao processar a requisição'}
def request_new_card_handler(event, context):
"""
POST - Esse código será chamado através de um POST para a API que será criada no arquivo serverless.
Faz a requisição de um novo cartão, o Score do candidato será avaliado.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi aprovado ou não.
"""
try:
body = event.get('body', {})
s3_bucket = getenv('S3_BUCKET', '')
if body:
# Verifica se o JSON é valido
json_valido, msg = valid_new_card_request(body)
if json_valido:
body['id'] = str(uuid.uuid4())
s3 = S3(s3_bucket)
s3.create_s3_instance()
# insere o JSON no S3
resp, msg = s3.put_s3_obj('', body, body['cpf'])
print(resp)
print(msg)
json_ret = {'status': 200, 'msg': 'Requisição enviada para aprovação!'}
else:
json_ret = {'status': 500, 'msg': msg}
else:
json_ret = {'status': 500, 'msg': 'Json inválido!'}
return json_ret
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def delete_card_request_handler(event, context):
"""
DELETE - Esse código será chamado através de um DELETE para a API que será criada no arquivo serverless.
Faz a requisição para apagar uma solicitação de cartão.
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: JSON contendo informações sobre a solicitação realizada, se o pedido de cartão foi excluído ou não.
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
del_response, msg = s3.delete_s3_obj('', path['id'])
if del_response['ResponseMetadata']['HTTPStatusCode'] == 204:
return {'status': 200, 'msg': 'Requisição deletada!'}
else:
return {'status': 404, 'msg': 'Requisição não encontrado'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
def process_card_request_handler(event, context):
"""
Realiza o processamento de um pedido de novo cartão.
Esse código é ativado quando um novo arquivo de pedido é salvo no S3
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: None.
"""
try:
for obj in event['Records']:
# Informações do arquivo que foi inserido no S3
bucket_name = obj['s3']['bucket']['name']
obj = obj['s3']['object']['key']
s3 = S3(bucket_name)
s3.create_s3_instance()
obj_json, msg = s3.get_s3_obj('', obj)
# faz os Score do cliente
score = randint(1, 999)
obj_json['credit'] = score_to_text(score, obj_json['income'])
resp, msg = s3.put_s3_obj('', obj_json, obj_json['cpf'])
except Exception as err:
print({'status': 500, 'msg': 'Erro interno ao processar a requisição', "error": f'{err}'})
def get_new_card_request_response_handler(event, context):
"""
GET - Esse código será cham
|
será criada no arquivo serverless.
Recebe o id do solicitante via parametro de URL e retorna as informações referente ao solicitante e seu crédito
:param event: Event recebido pela nuvem
:param context: Contexto com informações da função
:return: informações do solicitante
"""
try:
s3_bucket_crawler = getenv('S3_BUCKET', '')
print(event)
request_json, msg = None, None
path = event.get('path', {})
if 'id' in path.keys():
s3 = S3(s3_bucket_crawler)
s3.create_s3_instance()
request_json, msg = s3.get_s3_obj('', path['id'])
if request_json:
request_json['status'] = 200
return request_json
else:
return {'status': 404, 'msg': 'Requisição não encontrada'}
except Exception as err:
return {'status': 500, 'msg': 'Erro interno ao processar a requisição'}
if __name__ == '__main__':
ex_event = {'body':
{
'name': 'Leonardo Roberto Gazziro',
'phone': '99999999999',
'age': 26,
'cpf': '999.999.999-99',
'income': 3000
},
'method': 'POST', 'principalId': '',
'stage': 'dev',
'cognitoPoolClaims': {'sub': ''},
'enhancedAuthContext': {},
'headers': {'Accept': '*/*', 'Accept-Encoding': 'gzip, deflate, br',
'Cache-Control': 'no-cache', 'CloudFront-Forwarded-Proto': 'https',
'CloudFront-Is-Desktop-Viewer': 'true',
'CloudFront-Is-Mobile-Viewer': 'false',
'CloudFront-Is-SmartTV-Viewer': 'false',
'CloudFront-Is-Tablet-Viewer': 'false',
'CloudFront-Viewer-Country': 'BR',
'Content-Type': 'application/json',
'Host': '1nhucniq8b.execute-api.us-east-1.amazonaws.com',
'Postman-Token': '0d40ddc9-494c-438c-9e34-30f9f44ea018',
'User-Agent': 'PostmanRuntime/7.26.3',
'Via': '1.1 3fff6e22f8d6795a61bfdca17d362ca5.cloudfront.net (CloudFront)',
'X-Amz-Cf-Id': 'OKNx6jkzKLQ3nbtD0t4JTNynGlc2TZDSemAepsPC-8Kv0ZV1f6Tz7w==',
'X-Amzn-Trace-Id': 'Root=1-5f4bf435-7d8aa2621cebaaabfa719a0c',
'X-Forwarded-For': '138.204.24.213, 64.252.179.69',
'X-Forwarded-Port': '443', 'X-Forwarded-Proto': 'https'},
'query': {},
'path': {},
'identity': {'cognitoIdentityPoolId': '', 'accountId': '', 'cognitoIdentityId': '', 'caller': '',
'sourceIp': '138.204.24.213', 'principalOrgId': '', 'accessKey': '',
'cognitoAuthenticationType': '', 'cognitoAuthenticationProvider': '', 'userArn': '',
'userAgent': 'PostmanRuntime/7.26.3', 'user': ''},
'stageVariables': {},
'requestPath': ''}
ex_event_s3 = {'Records': [{'eventVersion': '2.1', 'eventSource': 'aws:s3', 'awsRegion': 'us-east-1',
'eventTime': '2020-09-06T18:11:57.722Z', 'eventName': 'ObjectCreated:Put',
'userIdentity': {'principalId': 'A1GN6EPM0JKA8K'},
'requestParameters': {'sourceIPAddress': '177.42.49.149'},
'responseElements': {'x-amz-request-id': '7857E87B7E4BA60D',
'x-amz-id-2': 'lmNysI3mKLmMQoOzCtPnjT8usl2fMUYbIyipfE59v3oWuyu44XxI/L2tXxRPkqjC6uUNu3rGB/eekMpW
|
ado através de um GET para a API que
|
identifier_name
|
config.go
|
created: %v", cfg)
return cfg
}
// EmptyConfig creates a Config with no services or groups
func EmptyConfig(workingDir string) Config {
log.Printf("Creating empty config\n")
cfg := Config{
workingDir: workingDir,
}
cfg.ServiceMap = make(map[string]*services.ServiceConfig)
cfg.GroupMap = make(map[string]*services.ServiceGroupConfig)
return cfg
}
// NormalizeServicePaths will modify the Paths for each of the provided services
// to be relative to the working directory of this config file
func (c *Config) NormalizeServicePaths(searchPath string, newServices []*services.ServiceConfig) ([]*services.ServiceConfig, error) {
log.Printf("Normalizing paths for %d services.\n", len(newServices))
var outServices []*services.ServiceConfig
for _, s := range newServices {
curService := *s
fullPath := filepath.Join(searchPath, *curService.Path)
relPath, err := filepath.Rel(c.workingDir, fullPath)
if err != nil {
return outServices, errors.WithStack(err)
}
curService.Path = &relPath
outServices = append(outServices, &curService)
}
return outServices, nil
}
// AppendServices adds services to an existing config without replacing existing services
func (c *Config) AppendServices(newServices []*services.ServiceConfig) error {
log.Printf("Appending %d services.\n", len(newServices))
if c.ServiceMap == nil {
c.ServiceMap = make(map[string]*services.ServiceConfig)
}
for _, s := range newServices {
if _, found := c.ServiceMap[s.Name]; !found {
c.ServiceMap[s.Name] = s
c.Services = append(c.Services, *s)
}
}
return nil
}
// AppendGroups adds groups to an existing config without replacing existing groups
func (c *Config) AppendGroups(groups []*services.ServiceGroupConfig) error {
var groupsDereferenced []services.ServiceGroupConfig
for _, group := range groups {
groupsDereferenced = append(groupsDereferenced, *group)
}
return errors.WithStack(c.AddGroups(groupsDereferenced))
}
func (c *Config) RemoveGroup(name string) error {
if _, ok := c.GroupMap[name]; !ok {
return errors.New("Group not found")
}
delete(c.GroupMap, name)
existingGroupDefs := c.Groups
c.Groups = make([]GroupDef, 0, len(existingGroupDefs))
for _, group := range existingGroupDefs {
if group.Name != name {
c.Groups = append(c.Groups, group)
}
}
return nil
}
// AddGroups adds a slice of groups to the Config
func (c *Config) AddGroups(groups []services.ServiceGroupConfig) error {
log.Printf("Adding %d groups.\n", len(groups))
for _, group := range groups {
grp := GroupDef{
Name: group.Name,
Aliases: group.Aliases,
Description: group.Description,
Children: []string{},
Env: group.Env,
}
for _, cg := range group.Groups {
if cg != nil {
grp.Children = append(grp.Children, cg.Name)
}
}
for _, cs := range group.Services {
if cs != nil {
grp.Children = append(grp.Children, cs.Name)
}
}
c.Groups = append(c.Groups, grp)
}
return nil
}
func (c *Config) loadImports() error {
log.Printf("Loading imports\n")
for _, i := range c.Imports {
var cPath string
if filepath.IsAbs(i) {
cPath = i
} else {
cPath = filepath.Join(c.workingDir, i)
}
log.Printf("Loading: %v\n", cPath)
r, err := os.Open(cPath)
if err != nil {
return errors.WithStack(err)
}
cfg, err := loadConfigContents(r, filepath.Dir(cPath))
if err != nil {
return errors.WithMessage(err, i)
}
err = c.importConfig(cfg)
if err != nil {
return errors.WithStack(err)
}
}
return nil
}
func (c *Config) importConfig(second Config) error {
for _, service := range append(second.Services, second.ImportedServices...) {
c.ImportedServices = append(c.ImportedServices, service)
}
for _, group := range append(second.Groups, second.ImportedGroups...) {
c.ImportedGroups = append(c.ImportedGroups, group)
}
return nil
}
func (c *Config) combinePath(path string) *string {
if filepath.IsAbs(path) || strings.HasPrefix(path, "$") {
return &path
}
fullPath := filepath.Join(c.workingDir, path)
return &fullPath
}
func addToMap(m map[string]struct{}, values ...string) {
for _, v := range values {
m[v] = struct{}{}
}
}
func intersect(m map[string]struct{}, values ...string) []string {
var out []string
for _, v := range values {
if _, ok := m[v]; ok {
out = append(out, v)
}
}
sort.Strings(out)
return out
}
func (c *Config) initMaps() error {
var err error
var svcs = make(map[string]*services.ServiceConfig)
var servicesSkipped = make(map[string]struct{})
var namesInUse = make(map[string]struct{})
for _, s := range append(c.Services, c.ImportedServices...) {
sc := s
sc.Env = append(sc.Env, c.Env...)
sc.ConfigFile, err = filepath.Abs(c.FilePath)
if err != nil {
return errors.WithStack(err)
}
if sc.MatchesPlatform() {
if i := intersect(namesInUse, append(sc.Aliases, sc.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join(i, ", "))
}
svcs[sc.Name] = &sc
addToMap(namesInUse, append(sc.Aliases, sc.Name)...)
} else {
servicesSkipped[sc.Name] = struct{}{}
}
}
var groups = make(map[string]*services.ServiceGroupConfig)
// First pass: Services
var orphanNames = make(map[string]struct{})
for _, g := range append(c.Groups, c.ImportedGroups...) {
var childServices []*services.ServiceConfig
for _, name := range g.Children {
if s, ok := svcs[name]; ok {
if s.Path != nil {
s.Path = c.combinePath(*s.Path)
}
childServices = append(childServices, s)
} else if _, skipped := servicesSkipped[name]; !skipped {
orphanNames[name] = struct{}{}
}
}
if i := intersect(namesInUse, append(g.Aliases, g.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join(i, ", "))
}
groups[g.Name] = &services.ServiceGroupConfig{
Name: g.Name,
Aliases: g.Aliases,
Description: g.Description,
Services: childServices,
Groups: []*services.ServiceGroupConfig{},
Env: g.Env,
ChildOrder: g.Children,
}
addToMap(namesInUse, append(g.Aliases, g.Name)...)
}
// Second pass: Groups
for _, g := range append(c.Groups, c.ImportedGroups...) {
childGroups := []*services.ServiceGroupConfig{}
for _, name := range g.Children {
if gr, ok := groups[name]; ok {
delete(orphanNames, name)
childGroups = append(childGroups, gr)
}
if hasChildCycle(groups[g.Name], childGroups) {
return errors.New("group cycle: " + g.Name)
}
}
groups[g.Name].Groups = childGroups
}
if len(orphanNames) > 0 {
var keys []string
for k := range orphanNames {
keys = append(keys, k)
}
return errors.New("A service or group could not be found for the following names: " + strings.Join(keys, ", "))
}
c.ServiceMap = svcs
c.GroupMap = groups
return nil
}
func hasChildCycle(parent *services.ServiceGroupConfig, children []*services.ServiceGroupConfig) bool {
for _, sg := range children {
if parent == sg {
return true
}
if hasChildCycle(parent, sg.Groups) {
return true
}
}
return false
}
func stringSliceIntersect(slices [][]string) []string {
var counts = make(map[string]int)
for _, s := range slices {
for _, v := range s {
counts[v]++
}
}
var outSlice []string
for v, count := range counts {
if count == len(slices) {
outSlice = append(outSlice, v)
}
}
return outSlice
}
func stringSliceRemoveCommon(common []string, original []string) []string {
var commonMap = make(map[string]interface{})
|
for _, s := range common {
commonMap[s] = struct{}{}
}
var outSlice []string
for _, s := range original {
|
random_line_split
|
|
config.go
|
)
if err != nil {
return Config{}, errors.WithStack(err)
}
workingDir := path.Dir(filePath)
config, err := loadConfigContents(reader, workingDir)
config.FilePath = filePath
if err != nil {
return Config{}, errors.WithStack(err)
}
if config.MinEdwardVersion != "" && edwardVersion != "" {
// Check that this config is supported by this version
minVersion, err1 := version.NewVersion(config.MinEdwardVersion)
if err1 != nil {
return Config{}, errors.WithStack(err)
}
currentVersion, err2 := version.NewVersion(edwardVersion)
if err2 != nil {
return Config{}, errors.WithStack(err)
}
if currentVersion.LessThan(minVersion) {
return Config{}, errors.New("this config requires at least version " + config.MinEdwardVersion)
}
}
err = config.initMaps()
log.Printf("Config loaded with: %d groups and %d services\n", len(config.GroupMap), len(config.ServiceMap))
return config, errors.WithStack(err)
}
// Reader from os.Open
func loadConfigContents(reader io.Reader, workingDir string) (Config, error) {
log.Printf("Loading config with working dir %v.\n", workingDir)
buf := new(bytes.Buffer)
_, err := buf.ReadFrom(reader)
if err != nil {
return Config{}, errors.Wrap(err, "could not read config")
}
data := buf.Bytes()
var config Config
err = json.Unmarshal(data, &config)
if err != nil {
if syntax, ok := err.(*json.SyntaxError); ok && syntax.Offset != 0 {
start := strings.LastIndex(string(data[:syntax.Offset]), "\n") + 1
line, pos := strings.Count(string(data[:start]), "\n")+1, int(syntax.Offset)-start-1
return Config{}, errors.Wrapf(err, "could not parse config file (line %v, char %v)", line, pos)
}
return Config{}, errors.Wrap(err, "could not parse config file")
}
config.workingDir = workingDir
err = config.loadImports()
if err != nil {
return Config{}, errors.WithStack(err)
}
return config, nil
}
// Save saves config to an io.Writer
func (c Config) Save(writer io.Writer) error {
log.Printf("Saving config")
content, err := json.MarshalIndent(c, "", " ")
if err != nil {
return errors.WithStack(err)
}
_, err = writer.Write(content)
return errors.WithStack(err)
}
// NewConfig creates a Config from slices of services and groups
func NewConfig(newServices []services.ServiceConfig, newGroups []services.ServiceGroupConfig) Config {
log.Printf("Creating new config with %d services and %d groups.\n", len(newServices), len(newGroups))
// Find Env settings common to all services
var allEnvSlices [][]string
for _, s := range newServices {
allEnvSlices = append(allEnvSlices, s.Env)
}
env := stringSliceIntersect(allEnvSlices)
// Remove common settings from services
var svcs []services.ServiceConfig
for _, s := range newServices {
s.Env = stringSliceRemoveCommon(env, s.Env)
svcs = append(svcs, s)
}
cfg := Config{
Env: env,
Services: svcs,
Groups: []GroupDef{},
}
cfg.AddGroups(newGroups)
log.Printf("Config created: %v", cfg)
return cfg
}
// EmptyConfig creates a Config with no services or groups
func EmptyConfig(workingDir string) Config {
log.Printf("Creating empty config\n")
cfg := Config{
workingDir: workingDir,
}
cfg.ServiceMap = make(map[string]*services.ServiceConfig)
cfg.GroupMap = make(map[string]*services.ServiceGroupConfig)
return cfg
}
// NormalizeServicePaths will modify the Paths for each of the provided services
// to be relative to the working directory of this config file
func (c *Config) NormalizeServicePaths(searchPath string, newServices []*services.ServiceConfig) ([]*services.ServiceConfig, error) {
log.Printf("Normalizing paths for %d services.\n", len(newServices))
var outServices []*services.ServiceConfig
for _, s := range newServices {
curService := *s
fullPath := filepath.Join(searchPath, *curService.Path)
relPath, err := filepath.Rel(c.workingDir, fullPath)
if err != nil {
return outServices, errors.WithStack(err)
}
curService.Path = &relPath
outServices = append(outServices, &curService)
}
return outServices, nil
}
// AppendServices adds services to an existing config without replacing existing services
func (c *Config) AppendServices(newServices []*services.ServiceConfig) error {
log.Printf("Appending %d services.\n", len(newServices))
if c.ServiceMap == nil {
c.ServiceMap = make(map[string]*services.ServiceConfig)
}
for _, s := range newServices {
if _, found := c.ServiceMap[s.Name]; !found {
c.ServiceMap[s.Name] = s
c.Services = append(c.Services, *s)
}
}
return nil
}
// AppendGroups adds groups to an existing config without replacing existing groups
func (c *Config) AppendGroups(groups []*services.ServiceGroupConfig) error {
var groupsDereferenced []services.ServiceGroupConfig
for _, group := range groups {
groupsDereferenced = append(groupsDereferenced, *group)
}
return errors.WithStack(c.AddGroups(groupsDereferenced))
}
func (c *Config) RemoveGroup(name string) error {
if _, ok := c.GroupMap[name]; !ok {
return errors.New("Group not found")
}
delete(c.GroupMap, name)
existingGroupDefs := c.Groups
c.Groups = make([]GroupDef, 0, len(existingGroupDefs))
for _, group := range existingGroupDefs {
if group.Name != name {
c.Groups = append(c.Groups, group)
}
}
return nil
}
// AddGroups adds a slice of groups to the Config
func (c *Config) AddGroups(groups []services.ServiceGroupConfig) error
|
c.Groups = append(c.Groups, grp)
}
return nil
}
func (c *Config) loadImports() error {
log.Printf("Loading imports\n")
for _, i := range c.Imports {
var cPath string
if filepath.IsAbs(i) {
cPath = i
} else {
cPath = filepath.Join(c.workingDir, i)
}
log.Printf("Loading: %v\n", cPath)
r, err := os.Open(cPath)
if err != nil {
return errors.WithStack(err)
}
cfg, err := loadConfigContents(r, filepath.Dir(cPath))
if err != nil {
return errors.WithMessage(err, i)
}
err = c.importConfig(cfg)
if err != nil {
return errors.WithStack(err)
}
}
return nil
}
func (c *Config) importConfig(second Config) error {
for _, service := range append(second.Services, second.ImportedServices...) {
c.ImportedServices = append(c.ImportedServices, service)
}
for _, group := range append(second.Groups, second.ImportedGroups...) {
c.ImportedGroups = append(c.ImportedGroups, group)
}
return nil
}
func (c *Config) combinePath(path string) *string {
if filepath.IsAbs(path) || strings.HasPrefix(path, "$") {
return &path
}
fullPath := filepath.Join(c.workingDir, path)
return &fullPath
}
func addToMap(m map[string]struct{}, values ...string) {
for _, v := range values {
m[v] = struct{}{}
}
}
func intersect(m map[string]struct{}, values ...string) []string {
var out []string
for _, v := range values {
if _, ok := m[v]; ok {
out = append(out, v)
}
}
sort.Strings(out)
return out
}
func (c *Config) initMaps() error {
var err error
var svcs = make(map[string]*services.ServiceConfig)
var servicesSkipped = make(map[string]struct{})
var namesInUse = make(map[string]struct{})
for _, s := range append(c.Services, c.ImportedServices...) {
sc := s
sc.Env = append(sc.Env, c.Env...)
sc.ConfigFile, err = filepath.Abs(c.FilePath)
if err != nil {
return errors.WithStack(err)
}
if sc.MatchesPlatform() {
if i := intersect(namesInUse, append(sc.Aliases, sc.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join
|
{
log.Printf("Adding %d groups.\n", len(groups))
for _, group := range groups {
grp := GroupDef{
Name: group.Name,
Aliases: group.Aliases,
Description: group.Description,
Children: []string{},
Env: group.Env,
}
for _, cg := range group.Groups {
if cg != nil {
grp.Children = append(grp.Children, cg.Name)
}
}
for _, cs := range group.Services {
if cs != nil {
grp.Children = append(grp.Children, cs.Name)
}
}
|
identifier_body
|
config.go
|
)
if err != nil {
return Config{}, errors.WithStack(err)
}
workingDir := path.Dir(filePath)
config, err := loadConfigContents(reader, workingDir)
config.FilePath = filePath
if err != nil {
return Config{}, errors.WithStack(err)
}
if config.MinEdwardVersion != "" && edwardVersion != "" {
// Check that this config is supported by this version
minVersion, err1 := version.NewVersion(config.MinEdwardVersion)
if err1 != nil {
return Config{}, errors.WithStack(err)
}
currentVersion, err2 := version.NewVersion(edwardVersion)
if err2 != nil {
return Config{}, errors.WithStack(err)
}
if currentVersion.LessThan(minVersion) {
return Config{}, errors.New("this config requires at least version " + config.MinEdwardVersion)
}
}
err = config.initMaps()
log.Printf("Config loaded with: %d groups and %d services\n", len(config.GroupMap), len(config.ServiceMap))
return config, errors.WithStack(err)
}
// Reader from os.Open
func loadConfigContents(reader io.Reader, workingDir string) (Config, error) {
log.Printf("Loading config with working dir %v.\n", workingDir)
buf := new(bytes.Buffer)
_, err := buf.ReadFrom(reader)
if err != nil {
return Config{}, errors.Wrap(err, "could not read config")
}
data := buf.Bytes()
var config Config
err = json.Unmarshal(data, &config)
if err != nil {
if syntax, ok := err.(*json.SyntaxError); ok && syntax.Offset != 0 {
start := strings.LastIndex(string(data[:syntax.Offset]), "\n") + 1
line, pos := strings.Count(string(data[:start]), "\n")+1, int(syntax.Offset)-start-1
return Config{}, errors.Wrapf(err, "could not parse config file (line %v, char %v)", line, pos)
}
return Config{}, errors.Wrap(err, "could not parse config file")
}
config.workingDir = workingDir
err = config.loadImports()
if err != nil {
return Config{}, errors.WithStack(err)
}
return config, nil
}
// Save saves config to an io.Writer
func (c Config) Save(writer io.Writer) error {
log.Printf("Saving config")
content, err := json.MarshalIndent(c, "", " ")
if err != nil {
return errors.WithStack(err)
}
_, err = writer.Write(content)
return errors.WithStack(err)
}
// NewConfig creates a Config from slices of services and groups
func NewConfig(newServices []services.ServiceConfig, newGroups []services.ServiceGroupConfig) Config {
log.Printf("Creating new config with %d services and %d groups.\n", len(newServices), len(newGroups))
// Find Env settings common to all services
var allEnvSlices [][]string
for _, s := range newServices {
allEnvSlices = append(allEnvSlices, s.Env)
}
env := stringSliceIntersect(allEnvSlices)
// Remove common settings from services
var svcs []services.ServiceConfig
for _, s := range newServices {
s.Env = stringSliceRemoveCommon(env, s.Env)
svcs = append(svcs, s)
}
cfg := Config{
Env: env,
Services: svcs,
Groups: []GroupDef{},
}
cfg.AddGroups(newGroups)
log.Printf("Config created: %v", cfg)
return cfg
}
// EmptyConfig creates a Config with no services or groups
func EmptyConfig(workingDir string) Config {
log.Printf("Creating empty config\n")
cfg := Config{
workingDir: workingDir,
}
cfg.ServiceMap = make(map[string]*services.ServiceConfig)
cfg.GroupMap = make(map[string]*services.ServiceGroupConfig)
return cfg
}
// NormalizeServicePaths will modify the Paths for each of the provided services
// to be relative to the working directory of this config file
func (c *Config) NormalizeServicePaths(searchPath string, newServices []*services.ServiceConfig) ([]*services.ServiceConfig, error) {
log.Printf("Normalizing paths for %d services.\n", len(newServices))
var outServices []*services.ServiceConfig
for _, s := range newServices {
curService := *s
fullPath := filepath.Join(searchPath, *curService.Path)
relPath, err := filepath.Rel(c.workingDir, fullPath)
if err != nil {
return outServices, errors.WithStack(err)
}
curService.Path = &relPath
outServices = append(outServices, &curService)
}
return outServices, nil
}
// AppendServices adds services to an existing config without replacing existing services
func (c *Config) AppendServices(newServices []*services.ServiceConfig) error {
log.Printf("Appending %d services.\n", len(newServices))
if c.ServiceMap == nil {
c.ServiceMap = make(map[string]*services.ServiceConfig)
}
for _, s := range newServices {
if _, found := c.ServiceMap[s.Name]; !found {
c.ServiceMap[s.Name] = s
c.Services = append(c.Services, *s)
}
}
return nil
}
// AppendGroups adds groups to an existing config without replacing existing groups
func (c *Config) AppendGroups(groups []*services.ServiceGroupConfig) error {
var groupsDereferenced []services.ServiceGroupConfig
for _, group := range groups {
groupsDereferenced = append(groupsDereferenced, *group)
}
return errors.WithStack(c.AddGroups(groupsDereferenced))
}
func (c *Config) RemoveGroup(name string) error {
if _, ok := c.GroupMap[name]; !ok {
return errors.New("Group not found")
}
delete(c.GroupMap, name)
existingGroupDefs := c.Groups
c.Groups = make([]GroupDef, 0, len(existingGroupDefs))
for _, group := range existingGroupDefs {
if group.Name != name {
c.Groups = append(c.Groups, group)
}
}
return nil
}
// AddGroups adds a slice of groups to the Config
func (c *Config)
|
(groups []services.ServiceGroupConfig) error {
log.Printf("Adding %d groups.\n", len(groups))
for _, group := range groups {
grp := GroupDef{
Name: group.Name,
Aliases: group.Aliases,
Description: group.Description,
Children: []string{},
Env: group.Env,
}
for _, cg := range group.Groups {
if cg != nil {
grp.Children = append(grp.Children, cg.Name)
}
}
for _, cs := range group.Services {
if cs != nil {
grp.Children = append(grp.Children, cs.Name)
}
}
c.Groups = append(c.Groups, grp)
}
return nil
}
func (c *Config) loadImports() error {
log.Printf("Loading imports\n")
for _, i := range c.Imports {
var cPath string
if filepath.IsAbs(i) {
cPath = i
} else {
cPath = filepath.Join(c.workingDir, i)
}
log.Printf("Loading: %v\n", cPath)
r, err := os.Open(cPath)
if err != nil {
return errors.WithStack(err)
}
cfg, err := loadConfigContents(r, filepath.Dir(cPath))
if err != nil {
return errors.WithMessage(err, i)
}
err = c.importConfig(cfg)
if err != nil {
return errors.WithStack(err)
}
}
return nil
}
func (c *Config) importConfig(second Config) error {
for _, service := range append(second.Services, second.ImportedServices...) {
c.ImportedServices = append(c.ImportedServices, service)
}
for _, group := range append(second.Groups, second.ImportedGroups...) {
c.ImportedGroups = append(c.ImportedGroups, group)
}
return nil
}
func (c *Config) combinePath(path string) *string {
if filepath.IsAbs(path) || strings.HasPrefix(path, "$") {
return &path
}
fullPath := filepath.Join(c.workingDir, path)
return &fullPath
}
func addToMap(m map[string]struct{}, values ...string) {
for _, v := range values {
m[v] = struct{}{}
}
}
func intersect(m map[string]struct{}, values ...string) []string {
var out []string
for _, v := range values {
if _, ok := m[v]; ok {
out = append(out, v)
}
}
sort.Strings(out)
return out
}
func (c *Config) initMaps() error {
var err error
var svcs = make(map[string]*services.ServiceConfig)
var servicesSkipped = make(map[string]struct{})
var namesInUse = make(map[string]struct{})
for _, s := range append(c.Services, c.ImportedServices...) {
sc := s
sc.Env = append(sc.Env, c.Env...)
sc.ConfigFile, err = filepath.Abs(c.FilePath)
if err != nil {
return errors.WithStack(err)
}
if sc.MatchesPlatform() {
if i := intersect(namesInUse, append(sc.Aliases, sc.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join
|
AddGroups
|
identifier_name
|
config.go
|
{
return errors.WithStack(err)
}
_, err = writer.Write(content)
return errors.WithStack(err)
}
// NewConfig creates a Config from slices of services and groups
func NewConfig(newServices []services.ServiceConfig, newGroups []services.ServiceGroupConfig) Config {
log.Printf("Creating new config with %d services and %d groups.\n", len(newServices), len(newGroups))
// Find Env settings common to all services
var allEnvSlices [][]string
for _, s := range newServices {
allEnvSlices = append(allEnvSlices, s.Env)
}
env := stringSliceIntersect(allEnvSlices)
// Remove common settings from services
var svcs []services.ServiceConfig
for _, s := range newServices {
s.Env = stringSliceRemoveCommon(env, s.Env)
svcs = append(svcs, s)
}
cfg := Config{
Env: env,
Services: svcs,
Groups: []GroupDef{},
}
cfg.AddGroups(newGroups)
log.Printf("Config created: %v", cfg)
return cfg
}
// EmptyConfig creates a Config with no services or groups
func EmptyConfig(workingDir string) Config {
log.Printf("Creating empty config\n")
cfg := Config{
workingDir: workingDir,
}
cfg.ServiceMap = make(map[string]*services.ServiceConfig)
cfg.GroupMap = make(map[string]*services.ServiceGroupConfig)
return cfg
}
// NormalizeServicePaths will modify the Paths for each of the provided services
// to be relative to the working directory of this config file
func (c *Config) NormalizeServicePaths(searchPath string, newServices []*services.ServiceConfig) ([]*services.ServiceConfig, error) {
log.Printf("Normalizing paths for %d services.\n", len(newServices))
var outServices []*services.ServiceConfig
for _, s := range newServices {
curService := *s
fullPath := filepath.Join(searchPath, *curService.Path)
relPath, err := filepath.Rel(c.workingDir, fullPath)
if err != nil {
return outServices, errors.WithStack(err)
}
curService.Path = &relPath
outServices = append(outServices, &curService)
}
return outServices, nil
}
// AppendServices adds services to an existing config without replacing existing services
func (c *Config) AppendServices(newServices []*services.ServiceConfig) error {
log.Printf("Appending %d services.\n", len(newServices))
if c.ServiceMap == nil {
c.ServiceMap = make(map[string]*services.ServiceConfig)
}
for _, s := range newServices {
if _, found := c.ServiceMap[s.Name]; !found {
c.ServiceMap[s.Name] = s
c.Services = append(c.Services, *s)
}
}
return nil
}
// AppendGroups adds groups to an existing config without replacing existing groups
func (c *Config) AppendGroups(groups []*services.ServiceGroupConfig) error {
var groupsDereferenced []services.ServiceGroupConfig
for _, group := range groups {
groupsDereferenced = append(groupsDereferenced, *group)
}
return errors.WithStack(c.AddGroups(groupsDereferenced))
}
func (c *Config) RemoveGroup(name string) error {
if _, ok := c.GroupMap[name]; !ok {
return errors.New("Group not found")
}
delete(c.GroupMap, name)
existingGroupDefs := c.Groups
c.Groups = make([]GroupDef, 0, len(existingGroupDefs))
for _, group := range existingGroupDefs {
if group.Name != name {
c.Groups = append(c.Groups, group)
}
}
return nil
}
// AddGroups adds a slice of groups to the Config
func (c *Config) AddGroups(groups []services.ServiceGroupConfig) error {
log.Printf("Adding %d groups.\n", len(groups))
for _, group := range groups {
grp := GroupDef{
Name: group.Name,
Aliases: group.Aliases,
Description: group.Description,
Children: []string{},
Env: group.Env,
}
for _, cg := range group.Groups {
if cg != nil {
grp.Children = append(grp.Children, cg.Name)
}
}
for _, cs := range group.Services {
if cs != nil {
grp.Children = append(grp.Children, cs.Name)
}
}
c.Groups = append(c.Groups, grp)
}
return nil
}
func (c *Config) loadImports() error {
log.Printf("Loading imports\n")
for _, i := range c.Imports {
var cPath string
if filepath.IsAbs(i) {
cPath = i
} else {
cPath = filepath.Join(c.workingDir, i)
}
log.Printf("Loading: %v\n", cPath)
r, err := os.Open(cPath)
if err != nil {
return errors.WithStack(err)
}
cfg, err := loadConfigContents(r, filepath.Dir(cPath))
if err != nil {
return errors.WithMessage(err, i)
}
err = c.importConfig(cfg)
if err != nil {
return errors.WithStack(err)
}
}
return nil
}
func (c *Config) importConfig(second Config) error {
for _, service := range append(second.Services, second.ImportedServices...) {
c.ImportedServices = append(c.ImportedServices, service)
}
for _, group := range append(second.Groups, second.ImportedGroups...) {
c.ImportedGroups = append(c.ImportedGroups, group)
}
return nil
}
func (c *Config) combinePath(path string) *string {
if filepath.IsAbs(path) || strings.HasPrefix(path, "$") {
return &path
}
fullPath := filepath.Join(c.workingDir, path)
return &fullPath
}
func addToMap(m map[string]struct{}, values ...string) {
for _, v := range values {
m[v] = struct{}{}
}
}
func intersect(m map[string]struct{}, values ...string) []string {
var out []string
for _, v := range values {
if _, ok := m[v]; ok {
out = append(out, v)
}
}
sort.Strings(out)
return out
}
func (c *Config) initMaps() error {
var err error
var svcs = make(map[string]*services.ServiceConfig)
var servicesSkipped = make(map[string]struct{})
var namesInUse = make(map[string]struct{})
for _, s := range append(c.Services, c.ImportedServices...) {
sc := s
sc.Env = append(sc.Env, c.Env...)
sc.ConfigFile, err = filepath.Abs(c.FilePath)
if err != nil {
return errors.WithStack(err)
}
if sc.MatchesPlatform() {
if i := intersect(namesInUse, append(sc.Aliases, sc.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join(i, ", "))
}
svcs[sc.Name] = &sc
addToMap(namesInUse, append(sc.Aliases, sc.Name)...)
} else {
servicesSkipped[sc.Name] = struct{}{}
}
}
var groups = make(map[string]*services.ServiceGroupConfig)
// First pass: Services
var orphanNames = make(map[string]struct{})
for _, g := range append(c.Groups, c.ImportedGroups...) {
var childServices []*services.ServiceConfig
for _, name := range g.Children {
if s, ok := svcs[name]; ok {
if s.Path != nil {
s.Path = c.combinePath(*s.Path)
}
childServices = append(childServices, s)
} else if _, skipped := servicesSkipped[name]; !skipped {
orphanNames[name] = struct{}{}
}
}
if i := intersect(namesInUse, append(g.Aliases, g.Name)...); len(i) > 0 {
return fmt.Errorf("Duplicate name or alias: %v", strings.Join(i, ", "))
}
groups[g.Name] = &services.ServiceGroupConfig{
Name: g.Name,
Aliases: g.Aliases,
Description: g.Description,
Services: childServices,
Groups: []*services.ServiceGroupConfig{},
Env: g.Env,
ChildOrder: g.Children,
}
addToMap(namesInUse, append(g.Aliases, g.Name)...)
}
// Second pass: Groups
for _, g := range append(c.Groups, c.ImportedGroups...) {
childGroups := []*services.ServiceGroupConfig{}
for _, name := range g.Children {
if gr, ok := groups[name]; ok {
delete(orphanNames, name)
childGroups = append(childGroups, gr)
}
if hasChildCycle(groups[g.Name], childGroups) {
return errors.New("group cycle: " + g.Name)
}
}
groups[g.Name].Groups = childGroups
}
if len(orphanNames) > 0
|
{
var keys []string
for k := range orphanNames {
keys = append(keys, k)
}
return errors.New("A service or group could not be found for the following names: " + strings.Join(keys, ", "))
}
|
conditional_block
|
|
scriptAppoint.js
|
;
case 9:
m = "10";
break;
case 10:
m = "11";
break;
case 11:
m = "12";
break;
}
fechaCita = a + "/" + m + "/" + d; //formato año/mes/dia hecho por evelyn
document.getElementById('fechaId').value = fechaCita;
//document.getElementById('lider').value = fechaCita;
return false;
}else {
window.alert("Por favor, verifica sí ya escogistes fecha de cita y horario");
return false;
}
return true;
}
// Mostrar todos los eventos de la fecha seleccionada en vistas de tarjeta
function show_events(events, month, day) {
// limpiar de datos los dos container
$(".events-container").empty(); /* */
// $(".events-container").show(250);
console.log(event_data["events"]);
// var date = event.data.date;
// Si no hay eventos para esta fecha, notifique al usuario
if (events.length === 0) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>There are no events planned for " + month + " " + day + ".</div>");
$(event_card).css({ "border-left": "10px solid #FF1744" });
$(event_card).append(event_name);
$(".events-container").append(event_card);
}
else {
// Go through and add each event as a card to the events container
for (var i = 0; i < events.length; i++) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>" + events[i]["occasion"] + ":</div>");
var event_count = $("<div class='event-count'>" + events[i]["invited_count"] + " Invited</div>");
if (events[i]["cancelled"] === true) {
$(event_card).css({
"border-left": "10px solid #FF1744"
});
event_count = $("<div class='event-cancelled'>Cancelled</div>");
}
$(event_card).append(event_name).append(event_count);
$(".events-container").append(event_card);
}
}
}
// Comprueba si una fecha específica tiene algún evento
function check_events(day, month, year) {
var events = [];
for (var i = 0; i < event_data["events"].length; i++) {
var event = event_data["events"][i];
if (event["day"] === day &&
event["month"] === month &&
event["year"] === year) {
events.push(event);
}
}
return events;
}
// // Datos dados para eventos en formato JSON
var event_data = {
"events": [
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Test Event",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 11
}
]
};
const months = [
"January",
"February",
"March",
"April",
"May",
"June",
"July",
"August",
"September",
"October",
"November",
"December"
];
// funcion para obtner la hora
function obtenerHor(id){
switch(id){
case "9":
document.getElementById('horaId').value = '09-10'; /* asignando un valor al id */
break;
case "10":
document.getElementById('horaId').value = '10-11';
ban2 = 1;
break;
case "11":
document.getElementById('horaId').value = '11-12';
ban2 = 1;
break;
case "12":
document.getElementById('horaId').value = "12-13";
ban2 = 1;
break
case "13":
document.getElementById('horaId').value = "13-14";
ban2 = 1;
break;
case "16":
document.getElementById('horaId').value = "16-17";
ban2 = 1;
break;
case "17":
document.getElementById('horaId').value = "17-18";
ban2 = 1;
break;
case "18":
document.getElementById('horaId').value = "18-19";
ban2 = 1;
break;
}
}
/* validacion de los campos del formmulario */
function validacionForm() {
var reason = "";
var nom = document.getElementById("name");
var mot = document.getElementById("cita");
var tel = document.getElementById("numer");
reason += validateName(nom);
reason += validateCita(mot);
reason += validatePhone(tel);
if (reason != "") {
window.alert("Algunos de los campos necesita correción\n" + reason);
return false;
}
return true;
}
/* validacion de la caja de nombre */
function validateName(nombre) {
var error = "";
// var illegalChars = /\W/; // permite letras, números y guiones bajos
if (nombre.value == "" || nombre.value == null || (/^\s+$/.test(nombre.value))) {
nombre.style.background = 'red';
error="La caja para nombre no contiene nada...\n";
nombre.focus();
}else if ((nombre.length < 3) || (nombre.length > 30)) {
nombre.style.background = 'red';
error = "El nombre tiene una longitud incorrecta...\n";
nombre.focus();
}else {
nombre.style.background = 'White';
}
// } else if (illegalChars.test(nombre.value)) {
// nombre.style.background = 'red';
// error = "El nombre ingresado contiene caracteres ilegales.\n";
// nombre.focus();
// }
return error;
}
function validateCita(cita){
var error = "";
if (cita.value == 0 || cita.length < 5 || cita.length > 30 ) {
cita.style.background = 'reed';
err
|
or = "Verifique el campo cita, antes de seguir\n"
} else {
cita.style.background = 'White';
|
conditional_block
|
|
scriptAppoint.js
|
0);
var date = event.data.date; // obtiene la fecha del ultimo dia del mes anterior
var mesAc = event.data.mes;// obtiene el mes actual, regresa un num
var new_month = $(".month").index(this); // regresa el nuemero del mes seleccionado
var ahioSel = date.getFullYear();
var ahioAc= event.data.anio;
if (mesAc==11 && ahioSel!==ahioAc){
if(new_month==0 || new_month==1){
$(".active-month").removeClass("active-month"); //remueve las clases
$(this).addClass("active-month"); // agrega una clase
date.setMonth(new_month); // se le asigna al dato al mes nuevo
init_calendar(date); //inicia el calendario con la fecha propor
}else{
window.alert("Mes no disponible");
}
} else if (new_month >= mesAc && new_month<= (mesAc + 2) && ahioSel==ahioAc) {
$(".active-month").removeClass("active-month"); //remueve las clases
$(this).addClass("active-month"); // agrega una clase
date.setMonth(new_month); // se le asigna al dato al mes nuevo
init_calendar(date); //inicia el calendario con la fecha propor
}
else {
window.alert("Mes no disponible");
}
}
//Controlador de eventos para cuando se hace clic en el botón derecho del año
function next_year(event, anio) {
$("#dialog").hide(250);
$(".events-container").hide(250);
var date = event.data.date; //obtiene la fecha del ultimo día del mes anterior
new_year = date.getFullYear() + 1; // devuelve el año actaul, y se le agrega 1 para el proximo año
if ((event.data.anio + 1) == new_year) {
$(".year").html(new_year); // se le agrega e texto a la clase
date.setFullYear(new_year); // le asigna al date, el año proximo con un mes antes
init_calendar(date);// el calendario se inicia con el ultimo dia del mes anterio
}
else{
window.alert("Año no disponible");
}
}
// Controlador de eventos para cuando se hace clic en el botón izquierdo del año
function prev_year(event, anio) {
$("#dialog").hide(250);
$(".events-container").hide(250);
var date = event.data.date; // obtiene la fecha del ultimo dia del mes anterior
new_year = date.getFullYear() - 1;// devuelve el año actaul, y se le le resta -1 para el año anterior
if (event.data.anio == new_year)
{
$("year").html(new_year);// se le agrega e texto a la clase
date.setFullYear(new_year);// le asigna al date, el año anterior, con fecha del ultimo dia del mes anterior a la que estamos
init_calendar(date);//inicia el calendario
}
}
// Controlador de eventos para hacer clic en el botón del nuevo evento
function new_event(event) {
if (ban==0){
window.alert("Antes de, escoge una fecha para tu cita.");
return false;
}
/*if (ban2 == 0) {
window.alert("Antes de, escoge una hora para tu cita.");
return false;
}*/
if(ban==1){
var date = event.data.date;
var a = date.getFullYear();
var m = date.getMonth();
$(".horarios-container").hide(250);
$(".events-container").hide(250);
$(".dialog").show(250); // muestra el contenedor para requisitar la cita
switch (m) {
case 0:
m = "01";
break;
case 1:
m = "02";
break;
case 2:
m = "03";
break;
case 3:
m = "04";
break;
case 4:
m = "05";
break;
case 5:
m = "06";
break;
case 6:
m = "07";
break;
case 7:
m = "08";
break;
case 8:
m = "09";
break;
case 9:
m = "10";
break;
case 10:
m = "11";
break;
case 11:
m = "12";
break;
}
fechaCita = a + "/" + m + "/" + d; //formato año/mes/dia hecho por evelyn
document.getElementById('fechaId').value = fechaCita;
//document.getElementById('lider').value = fechaCita;
return false;
}else {
window.alert("Por favor, verifica sí ya escogistes fecha de cita y horario");
return false;
}
return true;
}
// Mostrar todos los eventos de la fecha seleccionada en vistas de tarjeta
function show_events(events, month, day) {
// limpiar de datos los dos container
$(".events-container").empty(); /* */
// $(".events-container").show(250);
console.log(event_data["events"]);
// var date = event.data.date;
// Si no hay eventos para esta fecha, notifique al usuario
if (events.length === 0) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>There are no events planned for " + month + " " + day + ".</div>");
$(event_card).css({ "border-left": "10px solid #FF1744" });
$(event_card).append(event_name);
$(".events-container").append(event_card);
}
else {
// Go through and add each event as a card to the events container
for (var i = 0; i < events.length; i++) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>" + events[i]["occasion"] + ":</div>");
var event_count = $("<div class='event-count'>" + events[i]["invited_count"] + " Invited</div>");
if (events[i]["cancelled"] === true) {
$(event_card).css({
"border-left": "10px solid #FF1744"
});
event_count = $("<div class='event-cancelled'>Cancelled</div>");
}
$(event_card).append(event_name).append(event_count);
$(".events-container").append(event_card);
}
}
}
// Comprueba si una fecha específica tiene algún evento
function check_events(day, month, year) {
var events = [];
for (var i = 0; i < event_data["events"].length; i++) {
var event = event_data["events"][i];
if (event["day"] === day &&
event["month"] === month &&
event["year"] === year) {
events.push(event);
}
}
return events;
}
// // Datos dados para eventos en formato JSON
var event_data = {
"events": [
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
|
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
|
random_line_split
|
|
scriptAppoint.js
|
]["occasion"] + ":</div>");
var event_count = $("<div class='event-count'>" + events[i]["invited_count"] + " Invited</div>");
if (events[i]["cancelled"] === true) {
$(event_card).css({
"border-left": "10px solid #FF1744"
});
event_count = $("<div class='event-cancelled'>Cancelled</div>");
}
$(event_card).append(event_name).append(event_count);
$(".events-container").append(event_card);
}
}
}
// Comprueba si una fecha específica tiene algún evento
function check_events(day, month, year) {
var events = [];
for (var i = 0; i < event_data["events"].length; i++) {
var event = event_data["events"][i];
if (event["day"] === day &&
event["month"] === month &&
event["year"] === year) {
events.push(event);
}
}
return events;
}
// // Datos dados para eventos en formato JSON
var event_data = {
"events": [
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Test Event",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 11
}
]
};
const months = [
"January",
"February",
"March",
"April",
"May",
"June",
"July",
"August",
"September",
"October",
"November",
"December"
];
// funcion para obtner la hora
function obtenerHor(id){
switch(id){
case "9":
document.getElementById('horaId').value = '09-10'; /* asignando un valor al id */
break;
case "10":
document.getElementById('horaId').value = '10-11';
ban2 = 1;
break;
case "11":
document.getElementById('horaId').value = '11-12';
ban2 = 1;
break;
case "12":
document.getElementById('horaId').value = "12-13";
ban2 = 1;
break
case "13":
document.getElementById('horaId').value = "13-14";
ban2 = 1;
break;
case "16":
document.getElementById('horaId').value = "16-17";
ban2 = 1;
break;
case "17":
document.getElementById('horaId').value = "17-18";
ban2 = 1;
break;
case "18":
document.getElementById('horaId').value = "18-19";
ban2 = 1;
break;
}
}
/* validacion de los campos del formmulario */
function validacionForm() {
var reason = "";
var nom = document.getElementById("name");
var mot = document.getElementById("cita");
var tel = document.getElementById("numer");
reason += validateName(nom);
reason += validateCita(mot);
reason += validatePhone(tel);
if (reason != "") {
window.alert("Algunos de los campos necesita correción\n" + reason);
return false;
}
return true;
}
/* validacion de la caja de nombre */
function validateName(nombre) {
var error = "";
// var illegalChars = /\W/; // permite letras, números y guiones bajos
if (nombre.value == "" || nombre.value == null || (/^\s+$/.test(nombre.value))) {
nombre.style.background = 'red';
error="La caja para nombre no contiene nada...\n";
nombre.focus();
}else if ((nombre.length < 3) || (nombre.length > 30)) {
nombre.style.background = 'red';
error = "El nombre tiene una longitud incorrecta...\n";
nombre.focus();
}else {
nombre.style.background = 'White';
}
// } else if (illegalChars.test(nombre.value)) {
// nombre.style.background = 'red';
// error = "El nombre ingresado contiene caracteres ilegales.\n";
// nombre.focus();
// }
return error;
}
function validateCita(cita){
var error = "";
if (cita.value == 0 || cita.length < 5 || cita.length > 30 ) {
cita.style.background = 'reed';
error = "Verifique el campo cita, antes de seguir\n"
} else {
cita.style.background = 'White';
}
return error;
}
function validatePhone(tel) {
var error = "";
var stripped = tel.value.replace(/[\(\)\.\-\ ]/g, '');
if (tel.value == "" || tel.value==null) {
error = "No ingresó un número de teléfono..\n";
tel.style.background = 'red';
} else if (isNaN(parseInt(stripped))) {
error = "El número de teléfono contiene caracteres ilegales..\n";
tel.style.background = 'red';
} else if (!(stripped.length == 10)) {
error = "El número de teléfono tiene la longitud incorrecta. Asegúrese de incluir un código de área.\n";
tel.style.background = 'red';
}
return error;
}
// Obtener referencia a botones
// Recuerda: el punto . indica clases
const botones = document.querySelectorAll(".botonC");
// Definir función y evitar definirla de manera anónima
const cuandoSeHaceClick = function (evento) {
// Recuerda, this es el elemento
// window.alert("El texto que tiene es: ", this.innerText);
ban2=1;
}
// botones es un arreglo así que lo recorremos
botones.forEach(boton => {
//Agregar listener
boton.addEventListener("click", cuandoSeHaceClick);
});
function botonCancelar(){
$("#dialog").hide(250);
$(".horarios-container").s
|
how(250);
//$(".horariosss").show(250);
//$(".horariosss").show(250);
ban=0;
ban2=0;
document.getElementById('fechaId').innerHTML = "";
document.getElementById('horaId').innerHTML = "";
document.getElementById('name').innerHTML = "";
document.getElementById('cita').innerHTML = "";
document.getElementById('numer').innerHTML = "";
}
|
identifier_body
|
|
scriptAppoint.js
|
para tu cita.");
return false;
}
/*if (ban2 == 0) {
window.alert("Antes de, escoge una hora para tu cita.");
return false;
}*/
if(ban==1){
var date = event.data.date;
var a = date.getFullYear();
var m = date.getMonth();
$(".horarios-container").hide(250);
$(".events-container").hide(250);
$(".dialog").show(250); // muestra el contenedor para requisitar la cita
switch (m) {
case 0:
m = "01";
break;
case 1:
m = "02";
break;
case 2:
m = "03";
break;
case 3:
m = "04";
break;
case 4:
m = "05";
break;
case 5:
m = "06";
break;
case 6:
m = "07";
break;
case 7:
m = "08";
break;
case 8:
m = "09";
break;
case 9:
m = "10";
break;
case 10:
m = "11";
break;
case 11:
m = "12";
break;
}
fechaCita = a + "/" + m + "/" + d; //formato año/mes/dia hecho por evelyn
document.getElementById('fechaId').value = fechaCita;
//document.getElementById('lider').value = fechaCita;
return false;
}else {
window.alert("Por favor, verifica sí ya escogistes fecha de cita y horario");
return false;
}
return true;
}
// Mostrar todos los eventos de la fecha seleccionada en vistas de tarjeta
function show_events(events, month, day) {
// limpiar de datos los dos container
$(".events-container").empty(); /* */
// $(".events-container").show(250);
console.log(event_data["events"]);
// var date = event.data.date;
// Si no hay eventos para esta fecha, notifique al usuario
if (events.length === 0) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>There are no events planned for " + month + " " + day + ".</div>");
$(event_card).css({ "border-left": "10px solid #FF1744" });
$(event_card).append(event_name);
$(".events-container").append(event_card);
}
else {
// Go through and add each event as a card to the events container
for (var i = 0; i < events.length; i++) {
var event_card = $("<div class='event-card'></div>");
var event_name = $("<div class='event-name'>" + events[i]["occasion"] + ":</div>");
var event_count = $("<div class='event-count'>" + events[i]["invited_count"] + " Invited</div>");
if (events[i]["cancelled"] === true) {
$(event_card).css({
"border-left": "10px solid #FF1744"
});
event_count = $("<div class='event-cancelled'>Cancelled</div>");
}
$(event_card).append(event_name).append(event_count);
$(".events-container").append(event_card);
}
}
}
// Comprueba si una fecha específica tiene algún evento
function check_events(day, month, year) {
var events = [];
for (var i = 0; i < event_data["events"].length; i++) {
var event = event_data["events"][i];
if (event["day"] === day &&
event["month"] === month &&
event["year"] === year) {
events.push(event);
}
}
return events;
}
// // Datos dados para eventos en formato JSON
var event_data = {
"events": [
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10,
"cancelled": true
},
{
"occasion": " Repeated Test Event ",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 10
},
{
"occasion": " Test Event",
"invited_count": 120,
"year": 2017,
"month": 5,
"day": 11
}
]
};
const months = [
"January",
"February",
"March",
"April",
"May",
"June",
"July",
"August",
"September",
"October",
"November",
"December"
];
// funcion para obtner la hora
function obtenerHor(id){
switch(id){
case "9":
document.getElementById('horaId').value = '09-10'; /* asignando un valor al id */
break;
case "10":
document.getElementById('horaId').value = '10-11';
ban2 = 1;
break;
case "11":
document.getElementById('horaId').value = '11-12';
ban2 = 1;
break;
case "12":
document.getElementById('horaId').value = "12-13";
ban2 = 1;
break
case "13":
document.getElementById('horaId').value = "13-14";
ban2 = 1;
break;
case "16":
document.getElementById('horaId').value = "16-17";
ban2 = 1;
break;
case "17":
document.getElementById('horaId').value = "17-18";
ban2 = 1;
break;
case "18":
document.getElementById('horaId').value = "18-19";
ban2 = 1;
break;
}
}
/* validacion de los campos del formmulario */
function validacionForm() {
var reason = "";
var nom = document.getElementById("name");
var mot = document.getElementById("cita");
var tel = document.getElementById("numer");
reason += validateName(nom);
reason += validateCita(mot);
reason += validatePhone(tel);
if (reason != "") {
window.alert("Algunos de los campos necesita correción\n" + reason);
return false;
}
return true;
}
/* validacion de la caja de nombre */
function validateName(nombre) {
var error = ""
|
;
// var i
|
identifier_name
|
|
example_all.py
|
, 1.2, (100, 1)), axis=0)
train_y = objectiveFunction(train_x)
train_data = np.hstack([train_x, train_y])
test_x = np.linspace(-0.5, 1.5, 1000).reshape(-1, 1)
# test_x = np.sort(np.random.uniform(-1, 1, (100, 1)), axis=0)
test_y = objectiveFunction(test_x)
test_data = np.hstack([test_x, test_y])
train_x = train_data[:, :-1]
train_y = train_data[:, [-1]]
test_x = test_data[:, :-1]
test_y = test_data[:, [-1]]
N, D = train_x.shape
print("Dataset size: {} * {}".format(N, D))
if scoreMetric == "RMSE":
Nv = int(0.2 * N)
N -= Nv
val_x = train_x[-Nv:, :]
val_y = train_y[-Nv:].reshape(-1, 1)
train_x = train_x[:-Nv, :]
train_y = train_y[:-Nv].reshape(-1, 1)
vizData = False
if vizData:
plt.figure()
plt.plot(train_x, train_y, 'b-', label="train set")
if scoreMetric == "RMSE":
plt.plot(val_x, val_y, 'g-', label="validation set")
plt.plot(test_x, test_y, 'r-', label="test set")
plt.xlim(min(np.min(train_x), np.min(test_x)) - .2,
max(np.max(train_x), np.max(test_x)) + .2)
plt.ylim(min(np.min(train_y), np.min(test_y)) - .2,
max(np.max(train_y), np.max(test_y)) + .2)
plt.title("Data set")
plt.legend(loc=2)
plt.show()
exit()
"""
Quantile parametrisation
"""
if quantileParametrisation == parametrisations.BoundedQPoints:
qp = quantileParametrisation(6, p1=(0.03, -100), p2=(0.97, 100))
elif quantileParametrisation == parametrisations.StairCase:
qp = quantileParametrisation(9, p1=(0.03, -150), p2=(0.97, 150))
else:
qp = quantileParametrisation()
genParams = qp.gen_params
interpolator = qp.interpolator
paramLow = qp.paramLow
paramHigh = qp.paramHigh
scaleParams = qp.scale_params
if datasetName == "co2":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramHigh[0] = 2.3
paramLow[0] = 2
paramHigh[1] = 0.2
paramLow[1] = 0.01
elif datasetName == "airline":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramLow[0] = 2.3
paramHigh[0] = 2.8
paramLow[1] = 0.001
elif datasetName == "pores":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.45
elif datasetName == "rubber":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.25
elif datasetName == "tread":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.5
# paramLow[1] = 0.001
# paramHigh[1] = 0.01
if quantileParametrisation == parametrisations.InterpSingleSpline:
paramHigh[1] = 6
paramHigh[0] = 2.3
paramLow[0] = 1.5
"""
Kernel composition
"""
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
# composition = ["linear", "+", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
composition = ["bbq"]
if datasetName == "co2":
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
composition = ["linear_no_bias", "+", "bbq"]
elif datasetName == "airline":
composition = ["linear", "+", "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "linear_no_bias", "*", "bbq", "+", "bbq"]
if basicRBF:
# composition = ["rbf" if e == "bbq" else e for e in composition]
composition = ["rbf"]
qp.paramLow = np.array([-3])
qp.paramHigh = np.array([0])
qp.logScaleParams = np.array([True])
paramLow = qp.paramLow
paramHigh = qp.paramHigh
"""
Search/optimisation for best quantile
"""
if useARDkernel:
paramLow = np.array(D * list(paramLow))
paramHigh = np.array(D * list(paramHigh))
qp.logScaleParams = np.array(D * list(qp.logScaleParams))
if learnBeta:
paramLow = np.array([-3] + list(paramLow))
paramHigh = np.array([3] + list(paramHigh))
print("parameter space high: {}".format(paramHigh))
print("parameter space low: {}".format(paramLow))
score, allParams, bo = None, None, None
if searchStrategy == "Gridsearch":
assert (len(paramLow) == 2 and "Gridsearch only designed for 2 params")
X, Y = np.meshgrid(np.linspace(paramLow[0], paramHigh[0], gridRes),
np.linspace(paramLow[1], paramHigh[1], gridRes))
allParams = np.hstack([X.reshape(-1, 1), Y.reshape(-1, 1)])
score = np.zeros((gridRes ** 2,))
for i in tqdm(range(allParams.shape[0])):
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
elif searchStrategy == "BO":
search_int = np.vstack([np.atleast_2d(paramLow),
np.atleast_2d(paramHigh)]).T
bo = None # BO.BO(search_int, acq_fun=BO.UCB(boKappa), opt_maxeval=20,
allParams = np.zeros((nBOIter, len(paramLow)))
score = np.zeros((nBOIter,))
for i in tqdm(range(nBOIter)):
allParams[i, :] = bo.next_sample()
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
bo.update(allParams[i, :].reshape(1, -1),
np.array(score[i]).reshape(-1, 1))
elif searchStrategy == "NLopt":
allParams = np.zeros((nlOptIter, len(paramLow)))
score = np.zeros((nlOptIter,))
pbar = tqdm(total=nlOptIter)
i = 0
# Nlopt params
opt = nlopt.opt(nlOptAlgo, len(paramLow))
opt.set_lower_bounds(np.array(paramLow).reshape(-1))
opt.set_upper_bounds(np.array(paramHigh).reshape(-1))
opt.set_maxeval(nlOptIter)
def _fun_maximize(_x, grad):
|
global i
if i == nlOptIter:
print("Warning: maximum number of iterations reached.")
return float(np.min(score))
allParams[i, :] = _x
if learnBeta:
global blrBeta
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
curScore = float(-train_model(scaledParams, blrBeta))
# Keep track of previous tries
score[i] = curScore
i += 1
pbar.update(1)
|
identifier_body
|
|
example_all.py
|
= datasets.textures_2D(texture_name=datasetName)
else:
if datasetName == "cosine":
objectiveFunction = toy_functions.cosine
elif datasetName == "harmonics":
objectiveFunction = toy_functions.harmonics
elif datasetName == "pattern":
objectiveFunction = toy_functions.pattern
elif datasetName == "heaviside":
objectiveFunction = toy_functions.heaviside
elif datasetName == "quadraticCos":
objectiveFunction = toy_functions.quadratic_cos
elif datasetName == "steps":
objectiveFunction = toy_functions.steps
else:
raise RuntimeError("Objective function was not defined")
train_x = np.sort(np.random.uniform(-0.3, 1.2, (100, 1)), axis=0)
train_y = objectiveFunction(train_x)
train_data = np.hstack([train_x, train_y])
test_x = np.linspace(-0.5, 1.5, 1000).reshape(-1, 1)
# test_x = np.sort(np.random.uniform(-1, 1, (100, 1)), axis=0)
test_y = objectiveFunction(test_x)
test_data = np.hstack([test_x, test_y])
train_x = train_data[:, :-1]
train_y = train_data[:, [-1]]
test_x = test_data[:, :-1]
test_y = test_data[:, [-1]]
N, D = train_x.shape
print("Dataset size: {} * {}".format(N, D))
if scoreMetric == "RMSE":
Nv = int(0.2 * N)
N -= Nv
val_x = train_x[-Nv:, :]
val_y = train_y[-Nv:].reshape(-1, 1)
train_x = train_x[:-Nv, :]
train_y = train_y[:-Nv].reshape(-1, 1)
vizData = False
if vizData:
plt.figure()
plt.plot(train_x, train_y, 'b-', label="train set")
if scoreMetric == "RMSE":
plt.plot(val_x, val_y, 'g-', label="validation set")
plt.plot(test_x, test_y, 'r-', label="test set")
plt.xlim(min(np.min(train_x), np.min(test_x)) - .2,
max(np.max(train_x), np.max(test_x)) + .2)
plt.ylim(min(np.min(train_y), np.min(test_y)) - .2,
max(np.max(train_y), np.max(test_y)) + .2)
plt.title("Data set")
plt.legend(loc=2)
plt.show()
exit()
"""
Quantile parametrisation
"""
if quantileParametrisation == parametrisations.BoundedQPoints:
qp = quantileParametrisation(6, p1=(0.03, -100), p2=(0.97, 100))
elif quantileParametrisation == parametrisations.StairCase:
qp = quantileParametrisation(9, p1=(0.03, -150), p2=(0.97, 150))
else:
qp = quantileParametrisation()
genParams = qp.gen_params
interpolator = qp.interpolator
paramLow = qp.paramLow
paramHigh = qp.paramHigh
scaleParams = qp.scale_params
if datasetName == "co2":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramHigh[0] = 2.3
paramLow[0] = 2
paramHigh[1] = 0.2
paramLow[1] = 0.01
elif datasetName == "airline":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramLow[0] = 2.3
paramHigh[0] = 2.8
paramLow[1] = 0.001
elif datasetName == "pores":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.45
elif datasetName == "rubber":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.25
elif datasetName == "tread":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.5
# paramLow[1] = 0.001
# paramHigh[1] = 0.01
if quantileParametrisation == parametrisations.InterpSingleSpline:
paramHigh[1] = 6
paramHigh[0] = 2.3
paramLow[0] = 1.5
"""
Kernel composition
"""
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
# composition = ["linear", "+", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
composition = ["bbq"]
if datasetName == "co2":
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
composition = ["linear_no_bias", "+", "bbq"]
elif datasetName == "airline":
composition = ["linear", "+", "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "linear_no_bias", "*", "bbq", "+", "bbq"]
if basicRBF:
# composition = ["rbf" if e == "bbq" else e for e in composition]
composition = ["rbf"]
qp.paramLow = np.array([-3])
qp.paramHigh = np.array([0])
qp.logScaleParams = np.array([True])
paramLow = qp.paramLow
paramHigh = qp.paramHigh
"""
Search/optimisation for best quantile
"""
if useARDkernel:
paramLow = np.array(D * list(paramLow))
paramHigh = np.array(D * list(paramHigh))
qp.logScaleParams = np.array(D * list(qp.logScaleParams))
if learnBeta:
paramLow = np.array([-3] + list(paramLow))
paramHigh = np.array([3] + list(paramHigh))
print("parameter space high: {}".format(paramHigh))
print("parameter space low: {}".format(paramLow))
score, allParams, bo = None, None, None
if searchStrategy == "Gridsearch":
assert (len(paramLow) == 2 and "Gridsearch only designed for 2 params")
X, Y = np.meshgrid(np.linspace(paramLow[0], paramHigh[0], gridRes),
np.linspace(paramLow[1], paramHigh[1], gridRes))
allParams = np.hstack([X.reshape(-1, 1), Y.reshape(-1, 1)])
score = np.zeros((gridRes ** 2,))
for i in tqdm(range(allParams.shape[0])):
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
elif searchStrategy == "BO":
search_int = np.vstack([np.atleast_2d(paramLow),
np.atleast_2d(paramHigh)]).T
bo = None # BO.BO(search_int, acq_fun=BO.UCB(boKappa), opt_maxeval=20,
allParams = np.zeros((nBOIter, len(paramLow)))
score = np.zeros((nBOIter,))
for i in tqdm(range(nBOIter)):
allParams[i, :] = bo.next_sample()
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
bo.update(allParams[i, :].reshape(1, -1),
np.array(score[i]).reshape(-1, 1))
elif searchStrategy == "NLopt":
allParams = np.zeros((nlOptIter, len(paramLow)))
score = np.zeros((nlOptIter,))
pbar = tqdm(total=nlOptIter)
i = 0
# Nlopt params
opt = nlopt.opt(nlOptAlgo, len(paramLow))
opt.set_lower_bounds(np.array(paramLow).reshape(-1))
opt.set_upper_bounds(np.array(paramHigh).reshape(-1))
opt.set_maxeval(nlOptIter)
def
|
_fun_maximize
|
identifier_name
|
|
example_all.py
|
"linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "linear_no_bias", "*", "bbq", "+", "bbq"]
if basicRBF:
# composition = ["rbf" if e == "bbq" else e for e in composition]
composition = ["rbf"]
qp.paramLow = np.array([-3])
qp.paramHigh = np.array([0])
qp.logScaleParams = np.array([True])
paramLow = qp.paramLow
paramHigh = qp.paramHigh
"""
Search/optimisation for best quantile
"""
if useARDkernel:
paramLow = np.array(D * list(paramLow))
paramHigh = np.array(D * list(paramHigh))
qp.logScaleParams = np.array(D * list(qp.logScaleParams))
if learnBeta:
paramLow = np.array([-3] + list(paramLow))
paramHigh = np.array([3] + list(paramHigh))
print("parameter space high: {}".format(paramHigh))
print("parameter space low: {}".format(paramLow))
score, allParams, bo = None, None, None
if searchStrategy == "Gridsearch":
assert (len(paramLow) == 2 and "Gridsearch only designed for 2 params")
X, Y = np.meshgrid(np.linspace(paramLow[0], paramHigh[0], gridRes),
np.linspace(paramLow[1], paramHigh[1], gridRes))
allParams = np.hstack([X.reshape(-1, 1), Y.reshape(-1, 1)])
score = np.zeros((gridRes ** 2,))
for i in tqdm(range(allParams.shape[0])):
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
elif searchStrategy == "BO":
search_int = np.vstack([np.atleast_2d(paramLow),
np.atleast_2d(paramHigh)]).T
bo = None # BO.BO(search_int, acq_fun=BO.UCB(boKappa), opt_maxeval=20,
allParams = np.zeros((nBOIter, len(paramLow)))
score = np.zeros((nBOIter,))
for i in tqdm(range(nBOIter)):
allParams[i, :] = bo.next_sample()
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
bo.update(allParams[i, :].reshape(1, -1),
np.array(score[i]).reshape(-1, 1))
elif searchStrategy == "NLopt":
allParams = np.zeros((nlOptIter, len(paramLow)))
score = np.zeros((nlOptIter,))
pbar = tqdm(total=nlOptIter)
i = 0
# Nlopt params
opt = nlopt.opt(nlOptAlgo, len(paramLow))
opt.set_lower_bounds(np.array(paramLow).reshape(-1))
opt.set_upper_bounds(np.array(paramHigh).reshape(-1))
opt.set_maxeval(nlOptIter)
def _fun_maximize(_x, grad):
global i
if i == nlOptIter:
print("Warning: maximum number of iterations reached.")
return float(np.min(score))
allParams[i, :] = _x
if learnBeta:
global blrBeta
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
curScore = float(-train_model(scaledParams, blrBeta))
# Keep track of previous tries
score[i] = curScore
i += 1
pbar.update(1)
return curScore
opt.set_max_objective(_fun_maximize)
init_opt = np.random.uniform(0, 1, len(paramLow)) * \
(paramHigh - paramLow) + paramLow
opt.optimize(init_opt)
pbar.close()
"""
Saving
"""
# Compute best quantile
imax = np.argmax(score)
curParams = allParams[imax, :]
if learnBeta:
curParams = allParams[imax, 1:]
blrBeta = 10 ** allParams[imax, 0]
print("Best BLR beta: {}".format(blrBeta))
bestParam = scaleParams(curParams)
print("Best parameters: {}".format(bestParam))
bbq_qf = None
if not basicRBF:
# Interpolation
if useARDkernel:
bbq_qf = []
nprm = len(bestParam)
for j in range(D):
x, y, params = genParams(
bestParam[int(j * nprm / D):int((j + 1) * nprm / D)])
bbq_qf.append(InterpWithAsymptotes(x=x, y=y,
interpolator=interpolator,
params=params))
else:
x, y, params = genParams(bestParam)
bbq_qf = InterpWithAsymptotes(x=x, y=y, interpolator=interpolator,
params=params)
# Draw features with interpolated quantile
f_dict = {"linear_no_bias": ff.Linear(d=D, has_bias_term=False),
"linear": ff.Linear(d=D, has_bias_term=True)}
if basicRBF:
f_dict["rbf"] = ff.RFF_RBF(m=M, d=D, ls=bestParam)
else:
f_dict["bbq"] = ff.QMCF_BBQ(m=M, d=D,
sampler=partial(freqGenFn, bbq_qf=bbq_qf),
sequence_type=QMC_SEQUENCE.HALTON,
scramble_type=QMC_SCRAMBLING.GENERALISED,
qmc_kwargs={QMC_KWARG.PERM: None})
qmcf_bbq = ff.FComposition(f_dict=f_dict, composition=composition)
# Compute best data fit
blr = ABLR(k_phi=qmcf_bbq, d=D, d_out=1,
alpha=blrAlpha, beta=blrBeta, log_dir=None)
blr.learn_from_history(x_trn=train_x, y_trn=train_y)
prediction = blr.predict(x_tst=train_x, pred_var=True)
pred_train, predVar_train = prediction.mean, prediction.var
# Compute final score metrics
prediction = blr.predict(x_tst=test_x, pred_var=True)
pred_test, predVar_test = prediction.mean, prediction.var
finalRMSE = np.sqrt(np.mean((pred_test - test_y) ** 2))
finalMNLL = mnll(test_y, pred_test, predVar_test)
print("Final model RMSE: {}".format(finalRMSE))
print("Final model MNLL: {}".format(finalMNLL))
settingsVars = {
"RMSE": finalRMSE,
"MNLL": finalMNLL,
"bestParam": bestParam.tolist(),
"datasetName": datasetName,
"nlOptIter": nlOptIter,
"nlOptAlgo": nlOptAlgo,
"gridRes": gridRes,
"paramLow": paramLow.tolist(),
"paramHigh": paramHigh.tolist(),
"M": M,
"N": N,
"D": D,
"blrAlpha": blrAlpha,
"blrBeta": blrBeta,
"useARDkernel": useARDkernel,
"basicRBF": basicRBF,
"quantileParametrisation": quantileParametrisation.__name__,
"scoreMetric": scoreMetric,
"searchStrategy": searchStrategy,
"composition": composition,
"interpolator": interpolator.__name__
}
rs = ResultSaver(settingsVars)
rs.save_params_and_loss(allParams, -score)
if not basicRBF:
rs.save_quantile(bbq_qf, x, y)
rs.save_pdf(bbq_qf)
if D != 2:
if D == 1:
rs.save_dataset(train_x, train_y, test_x, test_y)
rs.save_data_fit(train_x, pred_train, predVar_train,
test_x, pred_test, predVar_test)
else:
# Generate images
imSize = 130
trainIdx = np.round((train_x + 1) * (imSize - 1) / 2.0).astype(int)
testIdx = np.round((test_x + 1) * (imSize - 1) / 2.0).astype(int)
fullImg = np.zeros((imSize, imSize))
for idx, y in zip(trainIdx, train_y):
fullImg[idx[0], idx[1]] = y
for idx, y in zip(testIdx, test_y):
fullImg[idx[0], idx[1]] = y
predImg = np.zeros((imSize, imSize))
for idx, y in zip(trainIdx, pred_train):
|
predImg[idx[0], idx[1]] = y
|
conditional_block
|
|
example_all.py
|
Algo = nlopt.GN_DIRECT
# nlOptAlgo = nlopt.GN_DIRECT_L
# nlOptAlgo = nlopt.GN_ISRES
# nlOptAlgo = nlopt.GN_ESCH
# NLopt Local algotithms (derivative free)
nlOptAlgo = nlopt.LN_COBYLA
# nlOptAlgo = nlopt.LN_BOBYQA
# nlOptAlgo = nlopt.LN_SBPLX
"""
Generate dataset
"""
if datasetName == "co2":
train_data, test_data = datasets.mauna_loa()
elif datasetName == "airline":
train_data, test_data = datasets.airline_passengers()
elif datasetName == "airfoil_noise":
train_data, test_data = datasets.airfoil_noise()
elif datasetName == "concrete":
train_data, test_data = datasets.concrete()
elif datasetName == "rubber" \
or datasetName == "pores" \
or datasetName == "tread":
train_data, test_data = datasets.textures_2D(texture_name=datasetName)
else:
if datasetName == "cosine":
objectiveFunction = toy_functions.cosine
elif datasetName == "harmonics":
objectiveFunction = toy_functions.harmonics
elif datasetName == "pattern":
objectiveFunction = toy_functions.pattern
elif datasetName == "heaviside":
objectiveFunction = toy_functions.heaviside
elif datasetName == "quadraticCos":
objectiveFunction = toy_functions.quadratic_cos
elif datasetName == "steps":
objectiveFunction = toy_functions.steps
else:
raise RuntimeError("Objective function was not defined")
train_x = np.sort(np.random.uniform(-0.3, 1.2, (100, 1)), axis=0)
train_y = objectiveFunction(train_x)
train_data = np.hstack([train_x, train_y])
test_x = np.linspace(-0.5, 1.5, 1000).reshape(-1, 1)
# test_x = np.sort(np.random.uniform(-1, 1, (100, 1)), axis=0)
test_y = objectiveFunction(test_x)
test_data = np.hstack([test_x, test_y])
train_x = train_data[:, :-1]
train_y = train_data[:, [-1]]
test_x = test_data[:, :-1]
test_y = test_data[:, [-1]]
N, D = train_x.shape
print("Dataset size: {} * {}".format(N, D))
if scoreMetric == "RMSE":
Nv = int(0.2 * N)
N -= Nv
val_x = train_x[-Nv:, :]
val_y = train_y[-Nv:].reshape(-1, 1)
train_x = train_x[:-Nv, :]
train_y = train_y[:-Nv].reshape(-1, 1)
vizData = False
if vizData:
plt.figure()
plt.plot(train_x, train_y, 'b-', label="train set")
if scoreMetric == "RMSE":
plt.plot(val_x, val_y, 'g-', label="validation set")
plt.plot(test_x, test_y, 'r-', label="test set")
plt.xlim(min(np.min(train_x), np.min(test_x)) - .2,
max(np.max(train_x), np.max(test_x)) + .2)
plt.ylim(min(np.min(train_y), np.min(test_y)) - .2,
max(np.max(train_y), np.max(test_y)) + .2)
plt.title("Data set")
plt.legend(loc=2)
plt.show()
exit()
"""
Quantile parametrisation
"""
if quantileParametrisation == parametrisations.BoundedQPoints:
qp = quantileParametrisation(6, p1=(0.03, -100), p2=(0.97, 100))
elif quantileParametrisation == parametrisations.StairCase:
qp = quantileParametrisation(9, p1=(0.03, -150), p2=(0.97, 150))
else:
qp = quantileParametrisation()
genParams = qp.gen_params
interpolator = qp.interpolator
paramLow = qp.paramLow
paramHigh = qp.paramHigh
scaleParams = qp.scale_params
if datasetName == "co2":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramHigh[0] = 2.3
paramLow[0] = 2
paramHigh[1] = 0.2
paramLow[1] = 0.01
elif datasetName == "airline":
if quantileParametrisation == parametrisations.PeriodicSimple:
paramLow[0] = 2.3
paramHigh[0] = 2.8
paramLow[1] = 0.001
elif datasetName == "pores":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.45
elif datasetName == "rubber":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.8
# paramLow[1] = 0.01
# paramHigh[1] = 0.25
elif datasetName == "tread":
pass
# paramLow[0] = 1.0
# paramHigh[0] = 2.5
# paramLow[1] = 0.001
# paramHigh[1] = 0.01
if quantileParametrisation == parametrisations.InterpSingleSpline:
paramHigh[1] = 6
paramHigh[0] = 2.3
paramLow[0] = 1.5
"""
Kernel composition
"""
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
# composition = ["linear", "+", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
# composition = ["linear", "+", "linear_no_bias", "*", "bbq"]
composition = ["bbq"]
if datasetName == "co2":
# composition = ["linear", "*", "linear_no_bias", "+", "bbq"]
composition = ["linear_no_bias", "+", "bbq"]
elif datasetName == "airline":
composition = ["linear", "+", "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "bbq"]
# composition = ["linear", "*", "linear_no_bias", "+",
# "linear", "*", "linear_no_bias", "*", "bbq", "+", "bbq"]
if basicRBF:
# composition = ["rbf" if e == "bbq" else e for e in composition]
composition = ["rbf"]
qp.paramLow = np.array([-3])
qp.paramHigh = np.array([0])
qp.logScaleParams = np.array([True])
|
"""
Search/optimisation for best quantile
"""
if useARDkernel:
paramLow = np.array(D * list(paramLow))
paramHigh = np.array(D * list(paramHigh))
qp.logScaleParams = np.array(D * list(qp.logScaleParams))
if learnBeta:
paramLow = np.array([-3] + list(paramLow))
paramHigh = np.array([3] + list(paramHigh))
print("parameter space high: {}".format(paramHigh))
print("parameter space low: {}".format(paramLow))
score, allParams, bo = None, None, None
if searchStrategy == "Gridsearch":
assert (len(paramLow) == 2 and "Gridsearch only designed for 2 params")
X, Y = np.meshgrid(np.linspace(paramLow[0], paramHigh[0], gridRes),
np.linspace(paramLow[1], paramHigh[1], gridRes))
allParams = np.hstack([X.reshape(-1, 1), Y.reshape(-1, 1)])
score = np.zeros((gridRes ** 2,))
for i in tqdm(range(allParams.shape[0])):
if learnBeta:
blrBeta = 10 ** (allParams[i, 0])
curParams = allParams[i, 1:]
else:
curParams = allParams[i, :]
scaledParams = scaleParams(curParams)
score[i] = -train_model(scaledParams, blrBeta)
elif searchStrategy == "BO":
search_int = np.vstack([np.atleast_2d(paramLow),
np.atleast_2d(paramHigh)]).T
bo = None # BO.BO(search_int, acq_fun=BO.UCB(boKappa), opt_maxeval=20,
allParams = np.zeros((nBOIter, len(paramLow)))
score = np.zeros((nBOIter,))
for i in tqdm(range(nBO
|
paramLow = qp.paramLow
paramHigh = qp.paramHigh
|
random_line_split
|
lib.rs
|
OneShot%3E%2C%20i16%2C%20CH%3E
//! [`set_data_rate()`]: struct.Ads1x1x.html#method.set_data_rate
//! [`set_full_scale_range()`]: struct.Ads1x1x.html#method.set_full_scale_range
//! [`is_measurement_in_progress()`]: struct.Ads1x1x.html#method.is_measurement_in_progress
//! [`set_high_threshold_raw()`]: struct.Ads1x1x.html#method.set_high_threshold_raw
//! [`set_comparator_mode()`]: struct.Ads1x1x.html#method.set_comparator_mode
//! [`set_comparator_polarity()`]: struct.Ads1x1x.html#method.set_comparator_polarity
//! [`set_comparator_latching()`]: struct.Ads1x1x.html#method.set_comparator_latching
//! [`set_comparator_queue()`]: struct.Ads1x1x.html#method.set_comparator_queue
//! [`disable_comparator()`]: struct.Ads1x1x.html#method.disable_comparator
//! [`use_alert_rdy_pin_as_ready()`]: struct.Ads1x1x.html#method.use_alert_rdy_pin_as_ready
//!
//! ## The devices
//!
//! The devices are precision, low power, 12/16-bit analog-to-digital
//! converters (ADC) that provide all features necessary to measure the most
//! common sensor signals in an ultra-small package. Depending on the device,
//! these integrate a programmable gain amplifier (PGA), voltage reference,
//! oscillator and high-accuracy temperature sensor.
//!
//! The devices can perform conversions at data rates up to 3300 samples per
//! second (SPS). The PGA offers input ranges from ±256 mV to ±6.144 V,
//! allowing both large and small signals to be measured with high resolution.
//! An input multiplexer (MUX) allows to measure two differential or four
//! single-ended inputs. The high-accuracy temperature sensor can be used for
//! system-level temperature monitoring or cold-junction compensation for
//! thermocouples.
//!
//! The devices operate either in continuous-conversion mode, or in a
//! single-shot mode that automatically powers down after a conversion.
//! Single-shot mode significantly reduces current consumption during idle
//! periods. Data is transferred through I2C.
//!
//! Here is a comparison of the caracteristics of the devices:
//!
//! | Device | Resolution | Sample Rate | Channels | Multi-channel | Features |
//! |---------|------------|--------------|----------|---------------|-----------------|
//! | ADS1013 | 12-bit | Max 3300 SPS | 1 | N/A | |
//! | ADS1014 | 12-bit | Max 3300 SPS | 1 | N/A | Comparator, PGA |
//! | ADS1015 | 12-bit | Max 3300 SPS | 4 | Multiplexed | Comparator, PGA |
//! | ADS1113 | 16-bit | Max 860 SPS | 1 | N/A | |
//! | ADS1114 | 16-bit | Max 860 SPS | 1 | N/A | Comparator, PGA |
//! | ADS1115 | 16-bit | Max 860 SPS | 4 | Multiplexed | Comparator, PGA |
//!
//! Datasheets:
//! - [ADS101x](http://www.ti.com/lit/ds/symlink/ads1015.pdf)
//! - [ADS111x](http://www.ti.com/lit/ds/symlink/ads1115.pdf)
//!
//! ## Usage examples (see also examples folder)
//!
//! To use this driver, import this crate and an `embedded_hal` implementation,
//! then instantiate the appropriate device.
//! In the following examples an instance of the device ADS1013 will be created
//! as an example. Other devices can be created with similar methods like:
//! `Ads1x1x::new_ads1114(...)`.
//!
//! Please find additional examples using hardware in this repository: [driver-examples]
//!
//! [driver-examples]: https://github.com/eldruin/driver-examples
//!
//! ### Create a driver instance for the ADS1013
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! // do something...
//!
//! // get the I2C device back
//! let dev = adc.destroy_ads1013();
//! ```
//!
//! ### Create a driver instance for the ADS1013 with an alternative address (method 1)
//!
|
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let (bit1, bit0) = (true, false); // last two bits of address
//! let address = SlaveAddr::Alternative(bit1, bit0);
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! ```
//! ### Create a driver instance for the ADS1013 with an alternative address (method 2)
//!
//! Using helper `SlaveAddr` creation method depending on the connection of
//! the `ADDR` pin.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! // `ADDR` pin connected to SDA results in the 0x4A effective address
//! let address = SlaveAddr::new_sda();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! ```
//!
//! ### Make a one-shot measurement
//! ```no_run
//! use ads1x1x::{channel, Ads1x1x, SlaveAddr};
//! use embedded_hal::adc::OneShot;
//! use linux_embedded_hal::I2cdev;
//! use nb::block;
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let mut adc = Ads1x1x::new_ads1013(dev, SlaveAddr::default());
//! let measurement = block!(adc.read(&mut channel::DifferentialA0A1)).unwrap();
//! println!("Measurement: {}", measurement);
//! let _dev = adc.destroy_ads1013(); // get I2C device back
//! ```
//!
//! ### Change into continuous conversion mode and read the last measurement
//!
//! Changing the mode may fail in case there was a communication error.
//! In this case, you can retrieve the unchanged device from the error type.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, ModeChangeError, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! match adc.into_continuous() {
//! Err(ModeChangeError::I2C(e, adc)) => /* mode change failed handling */ panic!(),
//! Ok(mut adc) => {
//! let measurement = adc.read().unwrap();
//! // ...
//! }
//! }
//! ```
//!
//!
//! ### Set the data rate
//! For 12-bit devices, the available data rates are given by `DataRate12Bit`.
//! For 16-bit devices, the available data rates are given by `DataRate16Bit`.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, DataRate16Bit, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let mut adc = Ads1x1x::new_ads1115(dev, address);
//! adc.set_data_rate(DataRate16Bit::Sps860).unwrap();
//! ```
//!
//! ### Configure the comparator
//! Configure the comparator to assert when the voltage drops below -1.5V
//! or goes above 1.5V in at least two consecutive conversions. Then the
//! ALERT/RDY pin will be set high and it will be kept so until the
//! measurement is read or an appropriate SMBus alert response is sent by
//! the master.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{
//! Ads1x1x, SlaveAddr, ComparatorQueue, ComparatorPolarity,
//! ComparatorMode, ComparatorLatching, FullScaleRange
//! };
//!
//! let dev = I2cdev::new("/dev/i2c-1
|
random_line_split
|
|
lib.rs
|
[`disable_comparator()`]: struct.Ads1x1x.html#method.disable_comparator
//! [`use_alert_rdy_pin_as_ready()`]: struct.Ads1x1x.html#method.use_alert_rdy_pin_as_ready
//!
//! ## The devices
//!
//! The devices are precision, low power, 12/16-bit analog-to-digital
//! converters (ADC) that provide all features necessary to measure the most
//! common sensor signals in an ultra-small package. Depending on the device,
//! these integrate a programmable gain amplifier (PGA), voltage reference,
//! oscillator and high-accuracy temperature sensor.
//!
//! The devices can perform conversions at data rates up to 3300 samples per
//! second (SPS). The PGA offers input ranges from ±256 mV to ±6.144 V,
//! allowing both large and small signals to be measured with high resolution.
//! An input multiplexer (MUX) allows to measure two differential or four
//! single-ended inputs. The high-accuracy temperature sensor can be used for
//! system-level temperature monitoring or cold-junction compensation for
//! thermocouples.
//!
//! The devices operate either in continuous-conversion mode, or in a
//! single-shot mode that automatically powers down after a conversion.
//! Single-shot mode significantly reduces current consumption during idle
//! periods. Data is transferred through I2C.
//!
//! Here is a comparison of the caracteristics of the devices:
//!
//! | Device | Resolution | Sample Rate | Channels | Multi-channel | Features |
//! |---------|------------|--------------|----------|---------------|-----------------|
//! | ADS1013 | 12-bit | Max 3300 SPS | 1 | N/A | |
//! | ADS1014 | 12-bit | Max 3300 SPS | 1 | N/A | Comparator, PGA |
//! | ADS1015 | 12-bit | Max 3300 SPS | 4 | Multiplexed | Comparator, PGA |
//! | ADS1113 | 16-bit | Max 860 SPS | 1 | N/A | |
//! | ADS1114 | 16-bit | Max 860 SPS | 1 | N/A | Comparator, PGA |
//! | ADS1115 | 16-bit | Max 860 SPS | 4 | Multiplexed | Comparator, PGA |
//!
//! Datasheets:
//! - [ADS101x](http://www.ti.com/lit/ds/symlink/ads1015.pdf)
//! - [ADS111x](http://www.ti.com/lit/ds/symlink/ads1115.pdf)
//!
//! ## Usage examples (see also examples folder)
//!
//! To use this driver, import this crate and an `embedded_hal` implementation,
//! then instantiate the appropriate device.
//! In the following examples an instance of the device ADS1013 will be created
//! as an example. Other devices can be created with similar methods like:
//! `Ads1x1x::new_ads1114(...)`.
//!
//! Please find additional examples using hardware in this repository: [driver-examples]
//!
//! [driver-examples]: https://github.com/eldruin/driver-examples
//!
//! ### Create a driver instance for the ADS1013
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! // do something...
//!
//! // get the I2C device back
//! let dev = adc.destroy_ads1013();
//! ```
//!
//! ### Create a driver instance for the ADS1013 with an alternative address (method 1)
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let (bit1, bit0) = (true, false); // last two bits of address
//! let address = SlaveAddr::Alternative(bit1, bit0);
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! ```
//! ### Create a driver instance for the ADS1013 with an alternative address (method 2)
//!
//! Using helper `SlaveAddr` creation method depending on the connection of
//! the `ADDR` pin.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! // `ADDR` pin connected to SDA results in the 0x4A effective address
//! let address = SlaveAddr::new_sda();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! ```
//!
//! ### Make a one-shot measurement
//! ```no_run
//! use ads1x1x::{channel, Ads1x1x, SlaveAddr};
//! use embedded_hal::adc::OneShot;
//! use linux_embedded_hal::I2cdev;
//! use nb::block;
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let mut adc = Ads1x1x::new_ads1013(dev, SlaveAddr::default());
//! let measurement = block!(adc.read(&mut channel::DifferentialA0A1)).unwrap();
//! println!("Measurement: {}", measurement);
//! let _dev = adc.destroy_ads1013(); // get I2C device back
//! ```
//!
//! ### Change into continuous conversion mode and read the last measurement
//!
//! Changing the mode may fail in case there was a communication error.
//! In this case, you can retrieve the unchanged device from the error type.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, ModeChangeError, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let adc = Ads1x1x::new_ads1013(dev, address);
//! match adc.into_continuous() {
//! Err(ModeChangeError::I2C(e, adc)) => /* mode change failed handling */ panic!(),
//! Ok(mut adc) => {
//! let measurement = adc.read().unwrap();
//! // ...
//! }
//! }
//! ```
//!
//!
//! ### Set the data rate
//! For 12-bit devices, the available data rates are given by `DataRate12Bit`.
//! For 16-bit devices, the available data rates are given by `DataRate16Bit`.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{Ads1x1x, DataRate16Bit, SlaveAddr};
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let mut adc = Ads1x1x::new_ads1115(dev, address);
//! adc.set_data_rate(DataRate16Bit::Sps860).unwrap();
//! ```
//!
//! ### Configure the comparator
//! Configure the comparator to assert when the voltage drops below -1.5V
//! or goes above 1.5V in at least two consecutive conversions. Then the
//! ALERT/RDY pin will be set high and it will be kept so until the
//! measurement is read or an appropriate SMBus alert response is sent by
//! the master.
//!
//! ```no_run
//! use linux_embedded_hal::I2cdev;
//! use ads1x1x::{
//! Ads1x1x, SlaveAddr, ComparatorQueue, ComparatorPolarity,
//! ComparatorMode, ComparatorLatching, FullScaleRange
//! };
//!
//! let dev = I2cdev::new("/dev/i2c-1").unwrap();
//! let address = SlaveAddr::default();
//! let mut adc = Ads1x1x::new_ads1015(dev, address);
//! adc.set_comparator_queue(ComparatorQueue::Two).unwrap();
//! adc.set_comparator_polarity(ComparatorPolarity::ActiveHigh).unwrap();
//! adc.set_comparator_mode(ComparatorMode::Window).unwrap();
//! adc.set_full_scale_range(FullScaleRange::Within2_048V).unwrap();
//! adc.set_low_threshold_raw(-1500).unwrap();
//! adc.set_high_threshold_raw(1500).unwrap();
//! adc.set_comparator_latching(ComparatorLatching::Latching).unwrap();
//! ```
#![doc(html_root_url = "https://docs.rs/ads1x1x/0.2.2")]
#![deny(unsafe_code)]
#![deny(missing_docs)]
#![no_std]
const DEVICE_BASE_ADDRESS: u8 = 0b100_1000;
struct Re
|
gister;
|
identifier_name
|
|
shear_calculation_withoutplot_vasp.py
|
3)) ####read basis vector from POSCAR
print(pos_array)
oa=np.array([pos_array[0][0], pos_array[0][1], pos_array[0][2]])
ob=np.array([pos_array[1][0], pos_array[1][1], pos_array[1][2]])
oc=np.array([pos_array[2][0], pos_array[2][1], pos_array[2][2]])
#------------------------------------------------------------------------------------
###start to calculate the structure data before rotation######
##a1_len=(float(pos_array[0][0])**2+float(pos_array[0][1])**2+float(pos_array[0][2])**2)**0.5
##a2_len=(float(pos_array[1][0])**2+float(pos_array[1][1])**2+float(pos_array[1][2])**2)**0.5
##a3_len=(float(pos_array[2][0])**2+float(pos_array[2][1])**2+float(pos_array[2][2])**2)**0.5
##
##P=np.array([P0[0]*a1_len,P0[1]*a2_len,P0[2]*a3_len])
##Q=np.array([Q0[0]*a1_len,Q0[1]*a2_len,Q0[2]*a3_len])
##
##XYZ_a=np.array([[a1_len, 0, 0],
## [0, a2_len, 0],
## [0, 0, a3_len]])
#----------------------------------------
basis=np.array([oa, ob, oc])
P0_abc=np.array([[P0[0], 0, 0],
[0, P0[1], 0],
[0, 0, P0[2]]])
P0_abc_basis=np.dot(P0_abc,basis)
P=P0_abc_basis[0]+P0_abc_basis[1]+P0_abc_basis[2] # the position of the normal of shear plane before rotation
Q0_abc=np.array([[Q0[0], 0, 0],
[0, Q0[1], 0],
[0, 0, Q0[2]]])
Q0_abc_basis=np.dot(Q0_abc,basis)
Q=Q0_abc_basis[0]+Q0_abc_basis[1]+Q0_abc_basis[2] # the position of the normal of shear plane after rotation
#----------------------------------------
#basis=np.array([oa, ob, oc])
L0_abc=np.array([[L0[0], 0, 0],
[0, L0[1], 0],
[0, 0, L0[2]]])
L0_abc_basis=np.dot(L0_abc,basis)
L=L0_abc_basis[0]+L0_abc_basis[1]+L0_abc_basis[2] # the position of shear direction before rotation
K0_abc=np.array([[K0[0], 0, 0],
[0, K0[1], 0],
[0, 0, K0[2]]])
K0_abc_basis=np.dot(K0_abc,basis)
K=K0_abc_basis[0]+K0_abc_basis[1]+K0_abc_basis[2] # the position of shear direction after rotation
#------------------------------------------------------------------------------------
###start to calculate the structure data after rotation######
############## the first rotation begin##############
######1.get the angel between two vectors########
P_norm=(P[0]**2+P[1]**2+P[2]**2)**0.5 #the norm of vetor before roation
Q_norm=(Q[0]**2+Q[1]**2+Q[2]**2)**0.5 #the norm of vetor after roation
PQ_dot=np.dot(P,Q)
theta=math.acos(PQ_dot/(P_norm*Q_norm))##obtain theta, the unit is radian instead of degree
theta_cos=PQ_dot/(P_norm*Q_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
M=np.array([P[1]*Q[2]-P[2]*Q[1], P[2]*Q[0]-P[0]*Q[2], P[0]*Q[1]-P[1]*Q[0]])
M_norm=(M[0]**2+M[1]**2+M[2]**2)**0.5 #the norm of M vetor
M_unit=np.array([M[0]/M_norm, M[1]/M_norm, M[2]/M_norm])
######3.get the rotation matrix###########
Nx=M_unit[0]
Ny=M_unit[1]
Nz=M_unit[2]
PQ_rot=np.array([[Nx*Nx*(1-theta_cos)+theta_cos, Nx*Ny*(1-theta_cos)+Nz*theta_sin, Nx*Nz*(1-theta_cos)-Ny*theta_sin],
[Nx*Ny*(1-theta_cos)-Nz*theta_sin, Ny*Ny*(1-theta_cos)+theta_cos, Ny*Nz*(1-theta_cos)+Nx*theta_sin],
[Nx*Nz*(1-theta_cos)+Ny*theta_sin, Ny*Nz*(1-theta_cos)-Nx*theta_sin, Nz*Nz*(1-theta_cos)+theta_cos]])
###### #####
a_PQ_rot=np.dot(pos_array,PQ_rot) #three basis vector after rotation
a_PQ_rot=a_PQ_rot.tolist()
L_rot=np.dot(L,PQ_rot)
############## the second rotation begin##############
######1.get the angel between two vectors########
L_norm=(L_rot[0]**2+L_rot[1]**2+L_rot[2]**2)**0.5 #the norm of vetor before roation
K_norm=(K[0]**2+K[1]**2+K[2]**2)**0.5 #the norm of vetor after roation
LK_dot=np.dot(L_rot,K)
theta=math.acos(LK_dot/(L_norm*K_norm))##obtain theta, the unit is radian instead of degree
theta_cos=LK_dot/(L_norm*K_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
H=np.array([L_rot[1]*K[2]-L_rot[2]*K[1], L_rot[2]*K[0]-L_rot[0]*K[2], L_rot[0]*K[1]-L_rot[1]*K[0]])
H_norm=(H[0]**2+H[1]**2+H[2]**2)**0.5 #the norm of M vetor
H_unit=np.array([H[0]/H_norm, H[1]/H_norm, H[2]/H_norm])
######3.get the rotation matrix###########
Ex=H_unit[0]
Ey=H_unit[1]
Ez=H_unit[2]
LK_rot=np.array([[Ex*Ex*(1-theta_cos)+theta_cos, Ex*Ey*(1-theta_cos)+Ez*theta_sin, Ex*Ez*(1-theta_cos)-Ey*theta_sin],
[Ex*Ey*(1-theta_cos)-Ez*theta_sin, Ey*Ey*(1-theta_cos)+theta_cos, Ey*Ez*(1-theta_cos)+Ex*theta_sin],
[Ex*Ez*(1-theta_cos)+Ey*theta_sin, Ey*Nz*(1-theta_cos)-Ex*theta_sin, Ez*Ez*(1-theta_cos)+theta_cos]])
###### #####
a_LK_rot=np.dot(a_PQ_rot,LK_rot) #three basis vector after rotation
a_LK_rot=a_LK_rot.tolist()
os.system("cp POSCAR POSCAR_original") #copy original POSCAR
r=[]
with open ("POSCAR") as poscar1:
for line in poscar1:
r.append(line)
f=open("POSCAR_rota","w") #write POSCAR after rotation as POSCAR_rota
for i in range(0,2):
f.write(r[i])
for j in range(len(a_LK_rot)):
f.write(str(a_LK_rot[j][0])+' ')
f.write(str(a_LK_rot[j][1])+' ')
f.write(str(a_LK_rot[j][2]))
f.write('\n')
for x in range(5,len(r)):
f.write(r[x])
f.close()
os.system("cp POSCAR_rota POSCAR") ## copy POSCAR_rota as POSCAR
###### Thirdly,apply tensile strain to the rotated basis vector
def postrain(poscar):
|
t2=[]
t3=[]
with open (poscar) as poscar2:
pos2=poscar2.readlines()
length=len(pos2)
for i in range(2,5):
j=pos2[i]
j=j.split()
t2.extend(j)
for i in range(len(t2)):
t3.extend([float(t2[i])])
pos_array3=np.array(t3).reshape((3,3))
if (Q0==np.array([1, 0, 0])).all():
stra_matr=np.array([[1,0,stra],
[0,1,0],
[stra,0,1]]) # strain(xz) and strain(zx)
a_stra=np.dot(pos_array3,stra_matr)
with open (poscar) as poscar:
s=[]
for line in poscar:
|
identifier_body
|
|
shear_calculation_withoutplot_vasp.py
|
, 0]
#P0=np.array([1, 1, 1]) #provide the axis before rotation.. P0 is the tensile direction we want to calculate.
###finish to read the input file
#Q0=np.array([1, 0, 0]) #after rotation. Q0 is the x axis. Since the x axis will not be optimized as we set in VASP, so Q0 is [1, 0, 0]
#------------------------------------------------------------------------------------
###open POSCAR
t=[]
t1=[]
with open ("POSCAR") as poscar0:
pos0=poscar0.readlines()
length=len(pos0)
for i in range(2,5):
j=pos0[i]
j=j.split()
t.extend(j)
print(t)
for i in range(len(t)):
t1.extend([float(t[i])])
pos_array=np.array(t1).reshape((3,3)) ####read basis vector from POSCAR
print(pos_array)
oa=np.array([pos_array[0][0], pos_array[0][1], pos_array[0][2]])
ob=np.array([pos_array[1][0], pos_array[1][1], pos_array[1][2]])
oc=np.array([pos_array[2][0], pos_array[2][1], pos_array[2][2]])
#------------------------------------------------------------------------------------
###start to calculate the structure data before rotation######
##a1_len=(float(pos_array[0][0])**2+float(pos_array[0][1])**2+float(pos_array[0][2])**2)**0.5
##a2_len=(float(pos_array[1][0])**2+float(pos_array[1][1])**2+float(pos_array[1][2])**2)**0.5
##a3_len=(float(pos_array[2][0])**2+float(pos_array[2][1])**2+float(pos_array[2][2])**2)**0.5
##
##P=np.array([P0[0]*a1_len,P0[1]*a2_len,P0[2]*a3_len])
##Q=np.array([Q0[0]*a1_len,Q0[1]*a2_len,Q0[2]*a3_len])
##
##XYZ_a=np.array([[a1_len, 0, 0],
## [0, a2_len, 0],
## [0, 0, a3_len]])
#----------------------------------------
basis=np.array([oa, ob, oc])
P0_abc=np.array([[P0[0], 0, 0],
[0, P0[1], 0],
[0, 0, P0[2]]])
P0_abc_basis=np.dot(P0_abc,basis)
P=P0_abc_basis[0]+P0_abc_basis[1]+P0_abc_basis[2] # the position of the normal of shear plane before rotation
Q0_abc=np.array([[Q0[0], 0, 0],
[0, Q0[1], 0],
[0, 0, Q0[2]]])
Q0_abc_basis=np.dot(Q0_abc,basis)
Q=Q0_abc_basis[0]+Q0_abc_basis[1]+Q0_abc_basis[2] # the position of the normal of shear plane after rotation
#----------------------------------------
#basis=np.array([oa, ob, oc])
L0_abc=np.array([[L0[0], 0, 0],
[0, L0[1], 0],
[0, 0, L0[2]]])
L0_abc_basis=np.dot(L0_abc,basis)
L=L0_abc_basis[0]+L0_abc_basis[1]+L0_abc_basis[2] # the position of shear direction before rotation
K0_abc=np.array([[K0[0], 0, 0],
[0, K0[1], 0],
[0, 0, K0[2]]])
K0_abc_basis=np.dot(K0_abc,basis)
K=K0_abc_basis[0]+K0_abc_basis[1]+K0_abc_basis[2] # the position of shear direction after rotation
#------------------------------------------------------------------------------------
###start to calculate the structure data after rotation######
############## the first rotation begin##############
######1.get the angel between two vectors########
P_norm=(P[0]**2+P[1]**2+P[2]**2)**0.5 #the norm of vetor before roation
Q_norm=(Q[0]**2+Q[1]**2+Q[2]**2)**0.5 #the norm of vetor after roation
PQ_dot=np.dot(P,Q)
theta=math.acos(PQ_dot/(P_norm*Q_norm))##obtain theta, the unit is radian instead of degree
theta_cos=PQ_dot/(P_norm*Q_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
M=np.array([P[1]*Q[2]-P[2]*Q[1], P[2]*Q[0]-P[0]*Q[2], P[0]*Q[1]-P[1]*Q[0]])
M_norm=(M[0]**2+M[1]**2+M[2]**2)**0.5 #the norm of M vetor
M_unit=np.array([M[0]/M_norm, M[1]/M_norm, M[2]/M_norm])
######3.get the rotation matrix###########
Nx=M_unit[0]
Ny=M_unit[1]
Nz=M_unit[2]
PQ_rot=np.array([[Nx*Nx*(1-theta_cos)+theta_cos, Nx*Ny*(1-theta_cos)+Nz*theta_sin, Nx*Nz*(1-theta_cos)-Ny*theta_sin],
[Nx*Ny*(1-theta_cos)-Nz*theta_sin, Ny*Ny*(1-theta_cos)+theta_cos, Ny*Nz*(1-theta_cos)+Nx*theta_sin],
[Nx*Nz*(1-theta_cos)+Ny*theta_sin, Ny*Nz*(1-theta_cos)-Nx*theta_sin, Nz*Nz*(1-theta_cos)+theta_cos]])
###### #####
a_PQ_rot=np.dot(pos_array,PQ_rot) #three basis vector after rotation
a_PQ_rot=a_PQ_rot.tolist()
L_rot=np.dot(L,PQ_rot)
############## the second rotation begin##############
######1.get the angel between two vectors########
L_norm=(L_rot[0]**2+L_rot[1]**2+L_rot[2]**2)**0.5 #the norm of vetor before roation
K_norm=(K[0]**2+K[1]**2+K[2]**2)**0.5 #the norm of vetor after roation
LK_dot=np.dot(L_rot,K)
theta=math.acos(LK_dot/(L_norm*K_norm))##obtain theta, the unit is radian instead of degree
theta_cos=LK_dot/(L_norm*K_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
H=np.array([L_rot[1]*K[2]-L_rot[2]*K[1], L_rot[2]*K[0]-L_rot[0]*K[2], L_rot[0]*K[1]-L_rot[1]*K[0]])
H_norm=(H[0]**2+H[1]**2+H[2]**2)**0.5 #the norm of M vetor
H_unit=np.array([H[0]/H_norm, H[1]/H_norm, H[2]/H_norm])
######3.get the rotation matrix###########
Ex=H_unit[0]
Ey=H_unit[1]
Ez=H_unit[2]
LK_rot=np.array([[Ex*Ex*(1-theta_cos)+theta_cos, Ex*Ey*(1-theta_cos)+Ez*theta_sin, Ex*Ez*(1-theta_cos)-Ey*theta_sin],
[Ex*Ey*(1-theta_cos)-Ez*theta_sin, Ey*Ey*(1-theta_cos)+theta_cos, Ey*Ez*(1-theta_cos)+Ex*theta_sin],
[Ex*Ez*(1-theta_cos)+Ey*theta_sin, Ey*Nz*(1-theta_cos)-Ex*theta_sin, Ez*Ez*(1-theta_cos)+theta_cos]])
###### #####
a_LK_rot=np.dot(a_PQ_rot,LK_rot) #three basis vector after rotation
a_LK_rot=a_LK_rot.tolist()
os.system("cp POSCAR POSCAR_original") #copy original POSCAR
r=[]
with open ("POSCAR") as poscar1:
for line in poscar1:
r.append(line)
f=open("POSCAR_rota","w") #write POSCAR after rotation as POSCAR_rota
for i in range(0,2):
f.write(r[i])
for j in range(len(a_LK_rot)):
f.write(str(a_LK_rot[j][0])+' ')
f.write(str(a_LK_rot[j][1])+' ')
f.write(str(a_LK_rot[j][2]))
f.write('\n')
for x in range(5,len(r)):
f.write(r[x])
f.close()
os.system("cp POSCAR_rota POSCAR") ## copy POSCAR_rota as POSCAR
###### Thirdly,apply tensile strain to the rotated basis vector
def
|
postrain
|
identifier_name
|
|
shear_calculation_withoutplot_vasp.py
|
_basis=np.dot(P0_abc,basis)
P=P0_abc_basis[0]+P0_abc_basis[1]+P0_abc_basis[2] # the position of the normal of shear plane before rotation
Q0_abc=np.array([[Q0[0], 0, 0],
[0, Q0[1], 0],
[0, 0, Q0[2]]])
Q0_abc_basis=np.dot(Q0_abc,basis)
Q=Q0_abc_basis[0]+Q0_abc_basis[1]+Q0_abc_basis[2] # the position of the normal of shear plane after rotation
#----------------------------------------
#basis=np.array([oa, ob, oc])
L0_abc=np.array([[L0[0], 0, 0],
[0, L0[1], 0],
[0, 0, L0[2]]])
L0_abc_basis=np.dot(L0_abc,basis)
L=L0_abc_basis[0]+L0_abc_basis[1]+L0_abc_basis[2] # the position of shear direction before rotation
K0_abc=np.array([[K0[0], 0, 0],
[0, K0[1], 0],
[0, 0, K0[2]]])
K0_abc_basis=np.dot(K0_abc,basis)
K=K0_abc_basis[0]+K0_abc_basis[1]+K0_abc_basis[2] # the position of shear direction after rotation
#------------------------------------------------------------------------------------
###start to calculate the structure data after rotation######
############## the first rotation begin##############
######1.get the angel between two vectors########
P_norm=(P[0]**2+P[1]**2+P[2]**2)**0.5 #the norm of vetor before roation
Q_norm=(Q[0]**2+Q[1]**2+Q[2]**2)**0.5 #the norm of vetor after roation
PQ_dot=np.dot(P,Q)
theta=math.acos(PQ_dot/(P_norm*Q_norm))##obtain theta, the unit is radian instead of degree
theta_cos=PQ_dot/(P_norm*Q_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
M=np.array([P[1]*Q[2]-P[2]*Q[1], P[2]*Q[0]-P[0]*Q[2], P[0]*Q[1]-P[1]*Q[0]])
M_norm=(M[0]**2+M[1]**2+M[2]**2)**0.5 #the norm of M vetor
M_unit=np.array([M[0]/M_norm, M[1]/M_norm, M[2]/M_norm])
######3.get the rotation matrix###########
Nx=M_unit[0]
Ny=M_unit[1]
Nz=M_unit[2]
PQ_rot=np.array([[Nx*Nx*(1-theta_cos)+theta_cos, Nx*Ny*(1-theta_cos)+Nz*theta_sin, Nx*Nz*(1-theta_cos)-Ny*theta_sin],
[Nx*Ny*(1-theta_cos)-Nz*theta_sin, Ny*Ny*(1-theta_cos)+theta_cos, Ny*Nz*(1-theta_cos)+Nx*theta_sin],
[Nx*Nz*(1-theta_cos)+Ny*theta_sin, Ny*Nz*(1-theta_cos)-Nx*theta_sin, Nz*Nz*(1-theta_cos)+theta_cos]])
###### #####
a_PQ_rot=np.dot(pos_array,PQ_rot) #three basis vector after rotation
a_PQ_rot=a_PQ_rot.tolist()
L_rot=np.dot(L,PQ_rot)
############## the second rotation begin##############
######1.get the angel between two vectors########
L_norm=(L_rot[0]**2+L_rot[1]**2+L_rot[2]**2)**0.5 #the norm of vetor before roation
K_norm=(K[0]**2+K[1]**2+K[2]**2)**0.5 #the norm of vetor after roation
LK_dot=np.dot(L_rot,K)
theta=math.acos(LK_dot/(L_norm*K_norm))##obtain theta, the unit is radian instead of degree
theta_cos=LK_dot/(L_norm*K_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
H=np.array([L_rot[1]*K[2]-L_rot[2]*K[1], L_rot[2]*K[0]-L_rot[0]*K[2], L_rot[0]*K[1]-L_rot[1]*K[0]])
H_norm=(H[0]**2+H[1]**2+H[2]**2)**0.5 #the norm of M vetor
H_unit=np.array([H[0]/H_norm, H[1]/H_norm, H[2]/H_norm])
######3.get the rotation matrix###########
Ex=H_unit[0]
Ey=H_unit[1]
Ez=H_unit[2]
LK_rot=np.array([[Ex*Ex*(1-theta_cos)+theta_cos, Ex*Ey*(1-theta_cos)+Ez*theta_sin, Ex*Ez*(1-theta_cos)-Ey*theta_sin],
[Ex*Ey*(1-theta_cos)-Ez*theta_sin, Ey*Ey*(1-theta_cos)+theta_cos, Ey*Ez*(1-theta_cos)+Ex*theta_sin],
[Ex*Ez*(1-theta_cos)+Ey*theta_sin, Ey*Nz*(1-theta_cos)-Ex*theta_sin, Ez*Ez*(1-theta_cos)+theta_cos]])
###### #####
a_LK_rot=np.dot(a_PQ_rot,LK_rot) #three basis vector after rotation
a_LK_rot=a_LK_rot.tolist()
os.system("cp POSCAR POSCAR_original") #copy original POSCAR
r=[]
with open ("POSCAR") as poscar1:
for line in poscar1:
r.append(line)
f=open("POSCAR_rota","w") #write POSCAR after rotation as POSCAR_rota
for i in range(0,2):
f.write(r[i])
for j in range(len(a_LK_rot)):
f.write(str(a_LK_rot[j][0])+' ')
f.write(str(a_LK_rot[j][1])+' ')
f.write(str(a_LK_rot[j][2]))
f.write('\n')
for x in range(5,len(r)):
f.write(r[x])
f.close()
os.system("cp POSCAR_rota POSCAR") ## copy POSCAR_rota as POSCAR
###### Thirdly,apply tensile strain to the rotated basis vector
def postrain(poscar):
t2=[]
t3=[]
with open (poscar) as poscar2:
pos2=poscar2.readlines()
length=len(pos2)
for i in range(2,5):
j=pos2[i]
j=j.split()
t2.extend(j)
for i in range(len(t2)):
t3.extend([float(t2[i])])
pos_array3=np.array(t3).reshape((3,3))
if (Q0==np.array([1, 0, 0])).all():
stra_matr=np.array([[1,0,stra],
[0,1,0],
[stra,0,1]]) # strain(xz) and strain(zx)
a_stra=np.dot(pos_array3,stra_matr)
with open (poscar) as poscar:
s=[]
for line in poscar:
s.append(line)
f=open("POSCAR_stra","w")
for i in range(0,2):
f.write(s[i])
for j in range(len(a_stra)):
f.write(str(a_stra[j][0])+' ')
f.write(str(a_stra[j][1])+' ')
f.write(str(a_stra[j][2]))
f.write('\n')
for x in range(5,len(s)):
f.write(s[x])
f.close()
###### Fourthly, performing VASP calculation
for i in np.arange(stra,stra*(times+1),stra):
|
print("****************************")
i=round(i,2)
print(i)
postrain("POSCAR")
os.system("cp POSCAR_stra POSCAR")
os.system("cp POSCAR_stra POSCAR_"+str(i))####
#os.system("bsub < Job_qsh.sh")
os.system("srun -n 20 /scratch/jin.zhang3_397857/Software/vasp.5.4.4/bin/vasp_std > vasp.out")
chek=os.popen("grep Voluntary OUTCAR").read()
while "Voluntary" not in chek:
chek=os.popen("grep Voluntary OUTCAR").read()
if "Voluntary" in chek:
break
os.system("cp OUTCAR OUTCAR_"+str(i))####
os.system("cp CONTCAR CONTCAR_"+str(i))####
os.system("grep 'in kB' OUTCAR > kB")
fin=os.popen("tail -1 kB").read()
fin=fin.split()
#os.system("cp CONTCAR POSCAR")
#os.system("bsub < Job_qsh.sh")
|
conditional_block
|
|
shear_calculation_withoutplot_vasp.py
|
1 -2 \n").split()]
if len(Sh_dir)==3:
Sh_dir=Sh_dir
elif len(Sh_pla)==4:
U=Sh_dir[1]+2*Sh_dir[0]
V=Sh_dir[0]+2*Sh_dir[1]
W=Sh_dir[3]
UVW_gcd=gcd(gcd(U,V),W)
Sh_dir=np.array([U/UVW_gcd, V/UVW_gcd, W/UVW_gcd])
L0=Sh_dir
K0=np.array([0, 0, 1]) #after rotation. K0 is the z axis. Since the x axis will not be optimized as we set in VASP, so Q0 is [1, 0, 0]
#P0=np.array([1, 1, 1]) #provide the axis before rotation.. P0 is the tensile direction we want to calculate.
###finish to read the input file
#Q0=np.array([1, 0, 0]) #after rotation. Q0 is the x axis. Since the x axis will not be optimized as we set in VASP, so Q0 is [1, 0, 0]
#------------------------------------------------------------------------------------
###open POSCAR
t=[]
t1=[]
with open ("POSCAR") as poscar0:
pos0=poscar0.readlines()
length=len(pos0)
for i in range(2,5):
j=pos0[i]
j=j.split()
t.extend(j)
print(t)
for i in range(len(t)):
t1.extend([float(t[i])])
pos_array=np.array(t1).reshape((3,3)) ####read basis vector from POSCAR
print(pos_array)
oa=np.array([pos_array[0][0], pos_array[0][1], pos_array[0][2]])
ob=np.array([pos_array[1][0], pos_array[1][1], pos_array[1][2]])
oc=np.array([pos_array[2][0], pos_array[2][1], pos_array[2][2]])
#------------------------------------------------------------------------------------
###start to calculate the structure data before rotation######
##a1_len=(float(pos_array[0][0])**2+float(pos_array[0][1])**2+float(pos_array[0][2])**2)**0.5
##a2_len=(float(pos_array[1][0])**2+float(pos_array[1][1])**2+float(pos_array[1][2])**2)**0.5
##a3_len=(float(pos_array[2][0])**2+float(pos_array[2][1])**2+float(pos_array[2][2])**2)**0.5
##
##P=np.array([P0[0]*a1_len,P0[1]*a2_len,P0[2]*a3_len])
##Q=np.array([Q0[0]*a1_len,Q0[1]*a2_len,Q0[2]*a3_len])
##
##XYZ_a=np.array([[a1_len, 0, 0],
## [0, a2_len, 0],
## [0, 0, a3_len]])
#----------------------------------------
basis=np.array([oa, ob, oc])
P0_abc=np.array([[P0[0], 0, 0],
[0, P0[1], 0],
[0, 0, P0[2]]])
P0_abc_basis=np.dot(P0_abc,basis)
P=P0_abc_basis[0]+P0_abc_basis[1]+P0_abc_basis[2] # the position of the normal of shear plane before rotation
Q0_abc=np.array([[Q0[0], 0, 0],
[0, Q0[1], 0],
[0, 0, Q0[2]]])
Q0_abc_basis=np.dot(Q0_abc,basis)
Q=Q0_abc_basis[0]+Q0_abc_basis[1]+Q0_abc_basis[2] # the position of the normal of shear plane after rotation
#----------------------------------------
#basis=np.array([oa, ob, oc])
L0_abc=np.array([[L0[0], 0, 0],
[0, L0[1], 0],
[0, 0, L0[2]]])
L0_abc_basis=np.dot(L0_abc,basis)
L=L0_abc_basis[0]+L0_abc_basis[1]+L0_abc_basis[2] # the position of shear direction before rotation
K0_abc=np.array([[K0[0], 0, 0],
[0, K0[1], 0],
[0, 0, K0[2]]])
K0_abc_basis=np.dot(K0_abc,basis)
K=K0_abc_basis[0]+K0_abc_basis[1]+K0_abc_basis[2] # the position of shear direction after rotation
#------------------------------------------------------------------------------------
###start to calculate the structure data after rotation######
############## the first rotation begin##############
######1.get the angel between two vectors########
P_norm=(P[0]**2+P[1]**2+P[2]**2)**0.5 #the norm of vetor before roation
Q_norm=(Q[0]**2+Q[1]**2+Q[2]**2)**0.5 #the norm of vetor after roation
PQ_dot=np.dot(P,Q)
theta=math.acos(PQ_dot/(P_norm*Q_norm))##obtain theta, the unit is radian instead of degree
theta_cos=PQ_dot/(P_norm*Q_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
M=np.array([P[1]*Q[2]-P[2]*Q[1], P[2]*Q[0]-P[0]*Q[2], P[0]*Q[1]-P[1]*Q[0]])
M_norm=(M[0]**2+M[1]**2+M[2]**2)**0.5 #the norm of M vetor
M_unit=np.array([M[0]/M_norm, M[1]/M_norm, M[2]/M_norm])
######3.get the rotation matrix###########
Nx=M_unit[0]
Ny=M_unit[1]
Nz=M_unit[2]
PQ_rot=np.array([[Nx*Nx*(1-theta_cos)+theta_cos, Nx*Ny*(1-theta_cos)+Nz*theta_sin, Nx*Nz*(1-theta_cos)-Ny*theta_sin],
[Nx*Ny*(1-theta_cos)-Nz*theta_sin, Ny*Ny*(1-theta_cos)+theta_cos, Ny*Nz*(1-theta_cos)+Nx*theta_sin],
[Nx*Nz*(1-theta_cos)+Ny*theta_sin, Ny*Nz*(1-theta_cos)-Nx*theta_sin, Nz*Nz*(1-theta_cos)+theta_cos]])
###### #####
a_PQ_rot=np.dot(pos_array,PQ_rot) #three basis vector after rotation
a_PQ_rot=a_PQ_rot.tolist()
L_rot=np.dot(L,PQ_rot)
############## the second rotation begin##############
######1.get the angel between two vectors########
L_norm=(L_rot[0]**2+L_rot[1]**2+L_rot[2]**2)**0.5 #the norm of vetor before roation
K_norm=(K[0]**2+K[1]**2+K[2]**2)**0.5 #the norm of vetor after roation
LK_dot=np.dot(L_rot,K)
theta=math.acos(LK_dot/(L_norm*K_norm))##obtain theta, the unit is radian instead of degree
theta_cos=LK_dot/(L_norm*K_norm)
theta_sin=math.sin(theta)
######2.get the rotation axis##############
H=np.array([L_rot[1]*K[2]-L_rot[2]*K[1], L_rot[2]*K[0]-L_rot[0]*K[2], L_rot[0]*K[1]-L_rot[1]*K[0]])
H_norm=(H[0]**2+H[1]**2+H[2]**2)**0.5 #the norm of M vetor
H_unit=np.array([H[0]/H_norm, H[1]/H_norm, H[2]/H_norm])
######3.get the rotation matrix###########
Ex=H_unit[0]
Ey=H_unit[1]
Ez=H_unit[2]
LK_rot=np.array([[Ex*Ex*(1-theta_cos)+theta_cos, Ex*Ey*(1-theta_cos)+Ez*theta_sin, Ex*Ez*(1-theta_cos)-Ey*theta_sin],
[Ex*Ey*(1-theta_cos)-Ez*theta_sin, Ey*Ey*(1-theta_cos)+theta_cos, Ey*Ez*(1-theta_cos)+Ex*theta_sin],
[Ex*Ez*(1-theta_cos)+Ey*theta_sin, Ey*Nz*(1-theta_cos)-Ex*theta_sin, Ez*Ez*(1-theta_cos)+theta_cos]])
###### #####
a_LK_rot=np.dot(a_PQ_rot,LK_rot) #three basis vector after rotation
a_LK_rot=a_LK_rot.tolist()
|
os.system("cp POSCAR POSCAR_original") #copy original POSCAR
r=[]
with open ("POSCAR") as poscar1:
|
random_line_split
|
|
lib.rs
|
Balance;
pub Locks get(fn locks): map T::AccountId => Vec<BalanceLock<T::Balance, T::Moment>>;
}
add_extra_genesis {
config(balances): Vec<(T::AccountId, T::Balance)>;
config(vesting): Vec<(T::AccountId, T::BlockNumber, T::BlockNumber, T::Balance)>;
// ^^ begin, length, amount liquid at genesis
}
}
decl_module! {
pub struct Module<T: Trait> for enum Call where origin: T::Origin {
fn deposit_event() = default;
/// Transfer some liquid free balance to another account.
///
/// `transfer` will set the `FreeBalance` of the sender and receiver.
/// It will decrease the total issuance of the system by the `TransferFee`.
/// If the sender's account is below the existential deposit as a result
/// of the transfer, the account will be reaped.
///
/// The dispatch origin for this call must be `Signed` by the transactor.
///
/// # <weight>
/// - Dependent on arguments but not critical, given proper implementations for
/// input config types. See related functions below.
/// - It contains a limited number of reads and writes internally and no complex computation.
///
/// Related functions:
///
/// - `ensure_can_withdraw` is always called internally but has a bounded complexity.
/// - Transferring balances to accounts that did not exist before will cause
/// `T::OnNewAccount::on_new_account` to be called.
/// - Removing enough funds from an account will trigger
/// `T::DustRemoval::on_unbalanced` and `T::OnFreeBalanceZero::on_free_balance_zero`.
/// - `transfer_keep_alive` works the same way as `transfer`, but has an additional
/// check that the transfer will not kill the origin account.
///
/// # </weight>
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn transfer(
origin,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
let transactor = ensure_signed(origin)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&transactor, &dest, value, ExistenceRequirement::AllowDeath)?;
}
/// Set the balances of a given account.
///
/// This will alter `FreeBalance` and `ReservedBalance` in storage. it will
/// also decrease the total issuance of the system (`TotalIssuance`).
/// If the new free or reserved balance is below the existential deposit,
/// it will reset the account nonce (`system::AccountNonce`).
///
/// The dispatch origin for this call is `root`.
///
/// # <weight>
/// - Independent of the arguments.
/// - Contains a limited number of reads and writes.
/// # </weight>
#[weight = SimpleDispatchInfo::FixedOperational(50_000)]
fn set_balance(
origin,
who: <T::Lookup as StaticLookup>::Source,
#[compact] new_free: T::Balance,
#[compact] new_reserved: T::Balance
) {
ensure_root(origin)?;
let who = T::Lookup::lookup(who)?;
let current_free = <FreeBalance<T>>::get(&who);
if new_free > current_free {
mem::drop(PositiveImbalance::<T>::new(new_free - current_free));
} else if new_free < current_free {
mem::drop(NegativeImbalance::<T>::new(current_free - new_free));
}
Self::set_free_balance(&who, new_free);
let current_reserved = <ReservedBalance<T>>::get(&who);
if new_reserved > current_reserved {
mem::drop(PositiveImbalance::<T>::new(new_reserved - current_reserved));
} else if new_reserved < current_reserved {
mem::drop(NegativeImbalance::<T>::new(current_reserved - new_reserved));
}
Self::set_reserved_balance(&who, new_reserved);
}
/// Exactly as `transfer`, except the origin must be root and the source account may be
/// specified.
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn force_transfer(
origin,
source: <T::Lookup as StaticLookup>::Source,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
ensure_root(origin)?;
let source = T::Lookup::lookup(source)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&source, &dest, value, ExistenceRequirement::AllowDeath)?;
}
}
}
impl<T: Trait> Module<T> {
// PUBLIC IMMUTABLES
/// Get the amount that is currently being vested and cannot be transferred out of this account.
pub fn vesting_balance(who: &T::AccountId) -> T::Balance {
if let Some(v) = Self::vesting(who) {
Self::free_balance(who).min(v.locked_at(<system::Module<T>>::block_number()))
} else {
Zero::zero()
}
}
// PRIVATE MUTABLES
/// Set the reserved balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_reserved_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
<ReservedBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
/// Set the free balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_free_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
// Commented out for now - but consider it instructive.
// assert!(!Self::total_balance(who).is_zero());
<FreeBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
}
// wrapping these imbalances in a private module is necessary to ensure absolute privacy
// of the inner member.
mod imbalances {
use rstd::mem;
use crate::{result, Imbalance, Saturating, StorageValue, Trait, Zero};
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been created without any equal and opposite accounting.
#[must_use]
pub struct PositiveImbalance<T: Trait>(T::Balance);
impl<T: Trait> PositiveImbalance<T> {
/// Create a new positive imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
PositiveImbalance(amount)
}
}
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been destroyed without any equal and opposite accounting.
#[must_use]
pub struct
|
<T: Trait>(T::Balance);
impl<T: Trait> NegativeImbalance<T> {
/// Create a new negative imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
NegativeImbalance(amount)
}
}
impl<T: Trait> Imbalance<T::Balance> for PositiveImbalance<T> {
type Opposite = NegativeImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else {
Err(self)
}
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite
|
NegativeImbalance
|
identifier_name
|
lib.rs
|
but has an additional
/// check that the transfer will not kill the origin account.
///
/// # </weight>
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn transfer(
origin,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
let transactor = ensure_signed(origin)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&transactor, &dest, value, ExistenceRequirement::AllowDeath)?;
}
/// Set the balances of a given account.
///
/// This will alter `FreeBalance` and `ReservedBalance` in storage. it will
/// also decrease the total issuance of the system (`TotalIssuance`).
/// If the new free or reserved balance is below the existential deposit,
/// it will reset the account nonce (`system::AccountNonce`).
///
/// The dispatch origin for this call is `root`.
///
/// # <weight>
/// - Independent of the arguments.
/// - Contains a limited number of reads and writes.
/// # </weight>
#[weight = SimpleDispatchInfo::FixedOperational(50_000)]
fn set_balance(
origin,
who: <T::Lookup as StaticLookup>::Source,
#[compact] new_free: T::Balance,
#[compact] new_reserved: T::Balance
) {
ensure_root(origin)?;
let who = T::Lookup::lookup(who)?;
let current_free = <FreeBalance<T>>::get(&who);
if new_free > current_free {
mem::drop(PositiveImbalance::<T>::new(new_free - current_free));
} else if new_free < current_free {
mem::drop(NegativeImbalance::<T>::new(current_free - new_free));
}
Self::set_free_balance(&who, new_free);
let current_reserved = <ReservedBalance<T>>::get(&who);
if new_reserved > current_reserved {
mem::drop(PositiveImbalance::<T>::new(new_reserved - current_reserved));
} else if new_reserved < current_reserved {
mem::drop(NegativeImbalance::<T>::new(current_reserved - new_reserved));
}
Self::set_reserved_balance(&who, new_reserved);
}
/// Exactly as `transfer`, except the origin must be root and the source account may be
/// specified.
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn force_transfer(
origin,
source: <T::Lookup as StaticLookup>::Source,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
ensure_root(origin)?;
let source = T::Lookup::lookup(source)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&source, &dest, value, ExistenceRequirement::AllowDeath)?;
}
}
}
impl<T: Trait> Module<T> {
// PUBLIC IMMUTABLES
/// Get the amount that is currently being vested and cannot be transferred out of this account.
pub fn vesting_balance(who: &T::AccountId) -> T::Balance {
if let Some(v) = Self::vesting(who) {
Self::free_balance(who).min(v.locked_at(<system::Module<T>>::block_number()))
} else {
Zero::zero()
}
}
// PRIVATE MUTABLES
/// Set the reserved balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_reserved_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
<ReservedBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
/// Set the free balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_free_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
// Commented out for now - but consider it instructive.
// assert!(!Self::total_balance(who).is_zero());
<FreeBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
}
// wrapping these imbalances in a private module is necessary to ensure absolute privacy
// of the inner member.
mod imbalances {
use rstd::mem;
use crate::{result, Imbalance, Saturating, StorageValue, Trait, Zero};
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been created without any equal and opposite accounting.
#[must_use]
pub struct PositiveImbalance<T: Trait>(T::Balance);
impl<T: Trait> PositiveImbalance<T> {
/// Create a new positive imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
PositiveImbalance(amount)
}
}
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been destroyed without any equal and opposite accounting.
#[must_use]
pub struct NegativeImbalance<T: Trait>(T::Balance);
impl<T: Trait> NegativeImbalance<T> {
/// Create a new negative imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
NegativeImbalance(amount)
}
}
impl<T: Trait> Imbalance<T::Balance> for PositiveImbalance<T> {
type Opposite = NegativeImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else {
Err(self)
}
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite> {
let (a, b) = (self.0, other.0);
mem::forget((self, other));
if a >= b {
Ok(Self(a - b))
} else {
Err(NegativeImbalance::new(b - a))
}
}
fn peek(&self) -> T::Balance {
self.0
}
}
impl<T: Trait> Imbalance<T::Balance> for NegativeImbalance<T> {
type Opposite = PositiveImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else {
Err(self)
}
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite> {
let (a, b) = (self.0, other.0);
mem::forget((self, other));
if a >= b {
Ok(Self(a - b))
} else {
Err(PositiveImbalance::new(b - a))
}
}
fn peek(&self) -> T::Balance {
self.0
}
}
|
random_line_split
|
||
lib.rs
|
{
mem::drop(NegativeImbalance::<T>::new(current_reserved - new_reserved));
}
Self::set_reserved_balance(&who, new_reserved);
}
/// Exactly as `transfer`, except the origin must be root and the source account may be
/// specified.
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn force_transfer(
origin,
source: <T::Lookup as StaticLookup>::Source,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
ensure_root(origin)?;
let source = T::Lookup::lookup(source)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&source, &dest, value, ExistenceRequirement::AllowDeath)?;
}
}
}
impl<T: Trait> Module<T> {
// PUBLIC IMMUTABLES
/// Get the amount that is currently being vested and cannot be transferred out of this account.
pub fn vesting_balance(who: &T::AccountId) -> T::Balance {
if let Some(v) = Self::vesting(who) {
Self::free_balance(who).min(v.locked_at(<system::Module<T>>::block_number()))
} else {
Zero::zero()
}
}
// PRIVATE MUTABLES
/// Set the reserved balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_reserved_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
<ReservedBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
/// Set the free balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_free_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
// Commented out for now - but consider it instructive.
// assert!(!Self::total_balance(who).is_zero());
<FreeBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
}
// wrapping these imbalances in a private module is necessary to ensure absolute privacy
// of the inner member.
mod imbalances {
use rstd::mem;
use crate::{result, Imbalance, Saturating, StorageValue, Trait, Zero};
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been created without any equal and opposite accounting.
#[must_use]
pub struct PositiveImbalance<T: Trait>(T::Balance);
impl<T: Trait> PositiveImbalance<T> {
/// Create a new positive imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
PositiveImbalance(amount)
}
}
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been destroyed without any equal and opposite accounting.
#[must_use]
pub struct NegativeImbalance<T: Trait>(T::Balance);
impl<T: Trait> NegativeImbalance<T> {
/// Create a new negative imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
NegativeImbalance(amount)
}
}
impl<T: Trait> Imbalance<T::Balance> for PositiveImbalance<T> {
type Opposite = NegativeImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else {
Err(self)
}
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite> {
let (a, b) = (self.0, other.0);
mem::forget((self, other));
if a >= b {
Ok(Self(a - b))
} else {
Err(NegativeImbalance::new(b - a))
}
}
fn peek(&self) -> T::Balance {
self.0
}
}
impl<T: Trait> Imbalance<T::Balance> for NegativeImbalance<T> {
type Opposite = PositiveImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else {
Err(self)
}
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite> {
let (a, b) = (self.0, other.0);
mem::forget((self, other));
if a >= b {
Ok(Self(a - b))
} else {
Err(PositiveImbalance::new(b - a))
}
}
fn peek(&self) -> T::Balance {
self.0
}
}
impl<T: Trait> Drop for PositiveImbalance<T> {
/// Basic drop handler will just square up the total issuance.
fn drop(&mut self) {
<super::TotalIssuance<T>>::mutate(|v| *v = v.saturating_add(self.0));
}
}
impl<T: Trait> Drop for NegativeImbalance<T> {
/// Basic drop handler will just square up the total issuance.
fn drop(&mut self) {
<super::TotalIssuance<T>>::mutate(|v| *v = v.saturating_sub(self.0));
}
}
}
impl<T: Trait> Currency<T::AccountId> for Module<T>
where
T::Balance: MaybeSerializeDeserialize + Debug,
{
type Balance = T::Balance;
type PositiveImbalance = PositiveImbalance<T>;
type NegativeImbalance = NegativeImbalance<T>;
fn total_balance(who: &T::AccountId) -> Self::Balance {
Self::free_balance(who) + Self::reserved_balance(who)
}
fn can_slash(who: &T::AccountId, value: Self::Balance) -> bool {
Self::free_balance(who) >= value
}
fn total_issuance() -> Self::Balance {
<TotalIssuance<T>>::get()
}
fn minimum_balance() -> Self::Balance {
Zero::zero()
}
fn burn(mut amount: Self::Balance) -> Self::PositiveImbalance {
<TotalIssuance<T>>::mutate(|issued| {
*issued = issued.checked_sub(&amount).unwrap_or_else(|| {
amount = *issued;
Zero::zero()
});
});
PositiveImbalance::new(amount)
}
fn issue(mut amount: Self::Balance) -> Self::NegativeImbalance {
<TotalIssuance<T>>::mutate(|issued| {
*issued = issued.checked_add(&amount).unwrap_or_else(|| {
amount = Self::Balance::max_value() - *issued;
Self::Balance::max_value()
})
});
NegativeImbalance::new(amount)
}
fn free_balance(who: &T::AccountId) -> Self::Balance
|
{
<FreeBalance<T>>::get(who)
}
|
identifier_body
|
|
lib.rs
|
;
pub Locks get(fn locks): map T::AccountId => Vec<BalanceLock<T::Balance, T::Moment>>;
}
add_extra_genesis {
config(balances): Vec<(T::AccountId, T::Balance)>;
config(vesting): Vec<(T::AccountId, T::BlockNumber, T::BlockNumber, T::Balance)>;
// ^^ begin, length, amount liquid at genesis
}
}
decl_module! {
pub struct Module<T: Trait> for enum Call where origin: T::Origin {
fn deposit_event() = default;
/// Transfer some liquid free balance to another account.
///
/// `transfer` will set the `FreeBalance` of the sender and receiver.
/// It will decrease the total issuance of the system by the `TransferFee`.
/// If the sender's account is below the existential deposit as a result
/// of the transfer, the account will be reaped.
///
/// The dispatch origin for this call must be `Signed` by the transactor.
///
/// # <weight>
/// - Dependent on arguments but not critical, given proper implementations for
/// input config types. See related functions below.
/// - It contains a limited number of reads and writes internally and no complex computation.
///
/// Related functions:
///
/// - `ensure_can_withdraw` is always called internally but has a bounded complexity.
/// - Transferring balances to accounts that did not exist before will cause
/// `T::OnNewAccount::on_new_account` to be called.
/// - Removing enough funds from an account will trigger
/// `T::DustRemoval::on_unbalanced` and `T::OnFreeBalanceZero::on_free_balance_zero`.
/// - `transfer_keep_alive` works the same way as `transfer`, but has an additional
/// check that the transfer will not kill the origin account.
///
/// # </weight>
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn transfer(
origin,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
let transactor = ensure_signed(origin)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&transactor, &dest, value, ExistenceRequirement::AllowDeath)?;
}
/// Set the balances of a given account.
///
/// This will alter `FreeBalance` and `ReservedBalance` in storage. it will
/// also decrease the total issuance of the system (`TotalIssuance`).
/// If the new free or reserved balance is below the existential deposit,
/// it will reset the account nonce (`system::AccountNonce`).
///
/// The dispatch origin for this call is `root`.
///
/// # <weight>
/// - Independent of the arguments.
/// - Contains a limited number of reads and writes.
/// # </weight>
#[weight = SimpleDispatchInfo::FixedOperational(50_000)]
fn set_balance(
origin,
who: <T::Lookup as StaticLookup>::Source,
#[compact] new_free: T::Balance,
#[compact] new_reserved: T::Balance
) {
ensure_root(origin)?;
let who = T::Lookup::lookup(who)?;
let current_free = <FreeBalance<T>>::get(&who);
if new_free > current_free {
mem::drop(PositiveImbalance::<T>::new(new_free - current_free));
} else if new_free < current_free {
mem::drop(NegativeImbalance::<T>::new(current_free - new_free));
}
Self::set_free_balance(&who, new_free);
let current_reserved = <ReservedBalance<T>>::get(&who);
if new_reserved > current_reserved {
mem::drop(PositiveImbalance::<T>::new(new_reserved - current_reserved));
} else if new_reserved < current_reserved {
mem::drop(NegativeImbalance::<T>::new(current_reserved - new_reserved));
}
Self::set_reserved_balance(&who, new_reserved);
}
/// Exactly as `transfer`, except the origin must be root and the source account may be
/// specified.
#[weight = SimpleDispatchInfo::FixedNormal(1_000_000)]
pub fn force_transfer(
origin,
source: <T::Lookup as StaticLookup>::Source,
dest: <T::Lookup as StaticLookup>::Source,
#[compact] value: T::Balance
) {
ensure_root(origin)?;
let source = T::Lookup::lookup(source)?;
let dest = T::Lookup::lookup(dest)?;
<Self as Currency<_>>::transfer(&source, &dest, value, ExistenceRequirement::AllowDeath)?;
}
}
}
impl<T: Trait> Module<T> {
// PUBLIC IMMUTABLES
/// Get the amount that is currently being vested and cannot be transferred out of this account.
pub fn vesting_balance(who: &T::AccountId) -> T::Balance {
if let Some(v) = Self::vesting(who) {
Self::free_balance(who).min(v.locked_at(<system::Module<T>>::block_number()))
} else {
Zero::zero()
}
}
// PRIVATE MUTABLES
/// Set the reserved balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_reserved_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
<ReservedBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
/// Set the free balance of an account to some new value. Will enforce `ExistentialDeposit`
/// law, annulling the account as needed.
///
/// Doesn't do any preparatory work for creating a new account, so should only be used when it
/// is known that the account already exists.
///
/// NOTE: LOW-LEVEL: This will not attempt to maintain total issuance. It is expected that
/// the caller will do this.
fn set_free_balance(who: &T::AccountId, balance: T::Balance) -> UpdateBalanceOutcome {
// Commented out for now - but consider it instructive.
// assert!(!Self::total_balance(who).is_zero());
<FreeBalance<T>>::insert(who, balance);
UpdateBalanceOutcome::Updated
}
}
// wrapping these imbalances in a private module is necessary to ensure absolute privacy
// of the inner member.
mod imbalances {
use rstd::mem;
use crate::{result, Imbalance, Saturating, StorageValue, Trait, Zero};
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been created without any equal and opposite accounting.
#[must_use]
pub struct PositiveImbalance<T: Trait>(T::Balance);
impl<T: Trait> PositiveImbalance<T> {
/// Create a new positive imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
PositiveImbalance(amount)
}
}
/// Opaque, move-only struct with private fields that serves as a token denoting that
/// funds have been destroyed without any equal and opposite accounting.
#[must_use]
pub struct NegativeImbalance<T: Trait>(T::Balance);
impl<T: Trait> NegativeImbalance<T> {
/// Create a new negative imbalance from a balance.
pub fn new(amount: T::Balance) -> Self {
NegativeImbalance(amount)
}
}
impl<T: Trait> Imbalance<T::Balance> for PositiveImbalance<T> {
type Opposite = NegativeImbalance<T>;
fn zero() -> Self {
Self(Zero::zero())
}
fn drop_zero(self) -> result::Result<(), Self> {
if self.0.is_zero() {
Ok(())
} else
|
}
fn split(self, amount: T::Balance) -> (Self, Self) {
let first = self.0.min(amount);
let second = self.0 - first;
mem::forget(self);
(Self(first), Self(second))
}
fn merge(mut self, other: Self) -> Self {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
self
}
fn subsume(&mut self, other: Self) {
self.0 = self.0.saturating_add(other.0);
mem::forget(other);
}
fn offset(self, other: Self::Opposite) -> result::Result<Self, Self::Opposite
|
{
Err(self)
}
|
conditional_block
|
query.py
|
链路是否主备线路,0表示备线路,1表示主线路
"over_drop", ##overlay非云网链路丢包率
"over_dropth", ##overlay非云网链路丢包率阈值
"over_sl_drop", ##overlay云网链路丢包率
"over_sl_dropth", ##overlay云网链路丢包率阈值
"over_delay", ##overlay链路时延
"over_shakedelay", ##overlay链路抖动
"over_mindelay",
"over_maxdelay",
"over_mdev",
"under_drop", ##underlay非云网链路丢包率
"under_dropth", ##underlay非云网链路丢包率阈值
"under_sl_drop", ##underlay云网链路丢包率
"under_sl_dropth", ##underlay云网链路丢包率阈值
"under_delay", ##underlay链路时延
"under_shakedelay", ##underlay链路抖动
"under_mindelay",
"under_maxdelay",
"under_mdev",
"tags",
#"metricset_name"
]
key2index = {k : i for i, k in enumerate(keys)}
def request_data(sess, postdata):
postdata = json.dumps(postdata)
response = sess.post(url, data=postdata)
res = response.text
data = json.loads(res)
samples = []
for i, a in enumerate(data['hits']['hits']):
pairs = a['_source']['pairs']
pair_dict = defaultdict(str)
pair_dict['timestamp'] = datetime.datetime.fromtimestamp(float(a['_source']['timestamp'])/1000)
if len(a['_source']['tags']) > 0:
pair_dict['tags'] = a['_source']['tags'][0]['value']
if len(a['_source']['tags']) > 1:
print('Warning:', a['_source']['tags'])
pair_dict['metricset_name'] = a['_source']['metricset']['name']
for p in pairs:
pair_dict[p['name']] = list(p['value'].values())[0]
samples.append([pair_dict[key] for key in keys])
return sorted(samples, key=lambda x:x[0]), data['hits']['total']
def get_list():
postdata = json.loads(json_query)
del postdata['query']['bool']['filter'][3:]
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
ta = datetime.datetime.now() - datetime.timedelta(hours=1)
tb = ta + datetime.timedelta(minutes=2)
timestamp_range['gte'] = ta.timestamp()*1000
timestamp_range['lt'] = tb.timestamp()*1000
logging.info('Request data in [{}, {})'.format(ta, tb))
sess = requests.Session()
sess.headers.update(headers)
for _ in range(3):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
break
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
df = pd.DataFrame(samples, columns=keys)
tunoip = {}
proname = {}
for nsid in df['ns_id'].unique():
if nsid=='': continue
t = df[df.ns_id==nsid]
tunoip[nsid] = sorted(list(t.tun_oip.unique()))
proname[nsid] = t.pop_proname.iloc[-1]
return tunoip, proname
def query(nsid, tunoip, begin_time=None, end_time=None, data_dir='./data'):
""" Query the ping data of specifical ns_id and tun_oip between specifical time range.
Time is datetime type or a string with format '%Y-%m-%d %H:%M:%S', e.g. '2019-10-13 10:09:00'
nsid: ns_id value.
tunoip: tun_oip value.
begin_time: The begining time to query. Default None means the time before 30 days.
end_time: The ending time to query. Default None means the current time.
"""
if type(begin_time) is str: begin_time = datetime.datetime.fromisoformat(begin_time)
if type(end_time) is str: end_time = datetime.datetime.fromisoformat(end_time)
nowtime = datetime.datetime.now()
pretime = nowtime - datetime.timedelta(days=30)
if end_time is None: end_time = nowtime
if begin_time is None: begin_time = pretime
q_begin_time, q_end_time = begin_time, end_time
if not os.path.exists(data_dir): os.mkdir(data_dir)
nsid_dir = os.path.join(data_dir, nsid)
if not os.path.exists(nsid_dir): os.mkdir(nsid_dir)
file_prefix = tunoip.replace('>', '')
# timestamp_path = os.path.join(nsid_dir, '%s.ts'%file_prefix)
csv_path = os.path.join(nsid_dir, '%s.csv'%file_prefix)
logging.info('Query: nsid={}, tunoip={}, timerange=[{}, {})'.format(nsid, tunoip, begin_time, end_time))
if os.path.exists(csv_path):
logging.info('Read data from %s'%csv_path)
df = pd.read_csv(csv_path)
logging.info('Time range: [{}, {}]'.format(df.iloc[0]['timestamp'], df.iloc[-1]['timestamp']))
begin_time = datetime.datetime.fromisoformat(df.iloc[-1]['timestamp']) + datetime.timedelta(minutes=1)
if begin_time < pretime:
begin_time = pretime
df.index = pd.to_datetime(df['timestamp'])
p_num = len(df)
if df.index[0] < pretime:
df = df.loc[pretime:]
logging.info('Forget %d records'%(p_num-len(df)))
else:
df = pd.DataFrame(columns=keys)
begin_time = pretime
# if nowtime - end_time < datetime.timedelta(hours=1):
# tmp_end_time = end_time - datetime.timedelta(hours=1)
# else:
# tmp_end_time = end_time
tmp_end_time = end_time
if q_end_time < begin_time:
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
return
s = json_query.replace('$nsid', nsid).replace('$tunoip', tunoip)
postdata = json.loads(s)
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
t_begin_time, t_end_time = int(begin_time.timestamp()), int(end_time.timestamp())
t_tmp_end_time = int(tmp_end_time.timestamp())
postdata['size'] = 10000
delta_time = 3600*24*3 # 7 days
sess = requests.Session()
sess.headers.update(headers)
def get_data(t, et):
if t >= et: return []
timestamp_range['gte'] = t*1000
timestamp_range['lt'] = et*1000
logging.info('Request data in [{}, {})'.format(datetime.datetime.fromtimestamp(t), datetime.datetime.fromtimestamp(et)))
for _ in range(5):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
return samples
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
data = []
for t in range(t_begin_time, t_tmp_end_time, delta_time):
et = min(t+delta_time, t_tmp_end_time)
data.extend(get_data(t, et))
df = df.append(pd.DataFrame(data, columns=keys), ignore_index=True)
# samples = get_data(t_tmp_end_time, t_end_time)
logging.info('Total Hists: {}'.format(len(data)))
if len(data)>0:
logging.info('Save data to %s'%csv_path)
df.to_csv(csv_path, index=False)
# df = df.append(pd.DataFrame(samples, columns=keys), ignore_index=True)
df.index = pd.to_datetime(df['timestamp'])
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
df.fillna(0, inplace=True)
return df
def cache():
tunoip, proname = get_list()
nsids = ['500850', '504608', '502902']
for nsid in nsids:
for lk in tunoip[nsid]:
query(nsid, lk, data_dir='/home/monitor/data')
# data_dir = '/home/monitor/data'
# fn = os.path.join(data_dir, 'collect.log')
fn = '/dev/stdout'
logging.basicConfig(level=logging.INFO,
format='[%(asctime)s] %(levelname)s: %(message)s',
datefmt='%Y-%m-%d %H:%M:%S')
|
random_line_split
|
||
query.py
|
filter": [
{"term": {"pairs.name": "tun_oip"}},
{"term": {"pairs.value.string": "$tunoip"}}
]
}
}
}
}
]
}
},
"size": 10000
}
'''
url = 'http://rpc.dsp.chinanetcenter.com:10200/api/console/proxy?path=*metricelf*%2F_search&method=POST'
headers = {
'Content-Type': 'application/json; charset=utf-8',
'Cookie': '__lcl=zh_CN; cluster=dashboard',
'DNT': '1',
'Host': 'rpc.dsp.chinanetcenter.com:10200',
'kbn-version': '6.1.1',
'Origin': 'http://rpc.dsp.chinanetcenter.com:10200',
'Referer': 'http://rpc.dsp.chinanetcenter.com:10200/app/kibana',
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/74.0.3729.108 Safari/537.36',
}
keys = [
"timestamp",
"pop_proname", ##项目名称
"pop_proid", ##oms上项目ID
"ns_id", ##项目的enterpriseID
"tun_uip", ##underlay链路ip
"tun_oip", ##overlay链路ip
"pop_ip", ##POP服务器管理ip
"tun_cost", ##链路权重
"tun_status", ##链路状态,0表示正常,1表示异常
"tun_cost_status", ##链路是否主备线路,0表示备线路,1表示主线路
"over_drop", ##overlay非云网链路丢包率
"over_dropth", ##overlay非云网链路丢包率阈值
"over_sl_drop", ##overlay云网链路丢包率
"over_sl_dropth", ##overlay云网链路丢包率阈值
"over_delay", ##overlay链路时延
"over_shakedelay", ##overlay链路抖动
"over_mindelay",
"over_maxdelay",
"over_mdev",
"under_drop", ##underlay非云网链路丢包率
"under_dropth", ##underlay非云网链路丢包率阈值
"under_sl_drop", ##underlay云网链路丢包率
"under_sl_dropth", ##underlay云网链路丢包率阈值
"under_delay", ##underlay链路时延
"under_shakedelay", ##underlay链路抖动
"under_mindelay",
"under_maxdelay",
"under_mdev",
"tags",
#"metricset_name"
]
key2index = {k : i for i, k in enumerate(keys)}
def request_data(sess, postdata):
postdata = json.dumps(postdata)
response = sess.post(url, data=postdata)
res = response.text
data = json.loads(res)
samples = []
for i, a in enumerate(data['hits']['hits']):
pairs = a['_source']['pairs']
pair_dict = defaultdict(str)
pair_dict['timestamp'] = datetime.datetime.fromtimestamp(float(a['_source']['timestamp'])/1000)
if len(a['_source']['tags']) > 0:
pair_dict['tags'] = a['_source']['tags'][0]['value']
if len(a['_source']['tags']) > 1:
print('Warning:', a['_source']['tags'])
pair_dict['metricset_name'] = a['_source']['metricset']['name']
for p in pairs:
pair_dict[p['name']] = list(p['value'].values())[0]
samples.append([pair_dict[key] for key in keys])
return sorted(samples, key=lambda x:x[0]), data['hits']['total']
def get_list():
postdata = json.loads(json_query)
del postdata['query']['bool']['filter'][3:]
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
ta = datetime.datetime.now() - datetime.timedelta(hours=1)
tb = ta + datetime.timedelta(minutes=2)
timestamp_range['gte'] = ta.timestamp()*1000
timestamp_range['lt'] = tb.timestamp()*1000
logging.info('Request data in [{}, {})'.format(ta, tb))
sess = requests.Session()
sess.headers.update(headers)
for _ in range(3):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
break
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
df = pd.DataFrame(samples, columns=keys)
tunoip = {}
proname = {}
for nsid in df['ns_id'].unique():
if nsid=='': continue
t = df[df.ns_id==nsid]
tunoip[nsid] = sorted(list(t.tun_oip.unique()))
proname[nsid] = t.pop_proname.iloc[-1]
return tunoip, proname
def query(nsid, tunoip, begin_time=None, end_time=None, data_dir='./data'):
""" Query the ping data of specifical ns_id and tun_oip between specifical time range.
Time is datetime type or a string with format '%Y-%m-%d %H:%M:%S', e.g. '2019-10-13 10:09:00'
nsid: ns_id value.
tunoip: tun_oip value.
begin_time: The begining time to query. Default None means the time before 30 days.
end_time: The ending time to query. Default None means the current time.
"""
if type(begin_time) is str: begin_time = datetime.datetime.fromisoformat(begin_time)
if type(end_time) is str: end_time = datetime.datetime.fromisoformat(end_time)
nowtime = datetime.datetime.now()
pretime = nowtime - datetime.timedelta(days=30)
if end_time is None: end_time = nowtime
if begin_time is None: begin_time = pretime
q_begin_time, q_end_time = begin_time, end_time
if not os.path.exists(data_dir): os.mkdir(data_dir)
nsid_dir = os.path.join(data_dir, nsid)
if not os.path.exists(nsid_dir): os.mkdir(nsid_dir)
file_prefix = tunoip.replace('>', '')
# timestamp_path = os.path.join(nsid_dir, '%s.ts'%file_prefix)
csv_path = os.path.join(nsid_dir, '%s.csv'%file_prefix)
logging.info('Query: nsid={}, tunoip={}, timerange=[{}, {})'.format(nsid, tunoip, begin_t
|
if os.path.exists(csv_path):
logging.info('Read data from %s'%csv_path)
df = pd.read_csv(csv_path)
logging.info('Time range: [{}, {}]'.format(df.iloc[0]['timestamp'], df.iloc[-1]['timestamp']))
begin_time = datetime.datetime.fromisoformat(df.iloc[-1]['timestamp']) + datetime.timedelta(minutes=1)
if begin_time < pretime:
begin_time = pretime
df.index = pd.to_datetime(df['timestamp'])
p_num = len(df)
if df.index[0] < pretime:
df = df.loc[pretime:]
logging.info('Forget %d records'%(p_num-len(df)))
else:
df = pd.DataFrame(columns=keys)
begin_time = pretime
# if nowtime - end_time < datetime.timedelta(hours=1):
# tmp_end_time = end_time - datetime.timedelta(hours=1)
# else:
# tmp_end_time = end_time
tmp_end_time = end_time
if q_end_time < begin_time:
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
return
s = json_query.replace('$nsid', nsid).replace('$tunoip', tunoip)
postdata = json.loads(s)
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
t_begin_time, t_end_time = int(begin_time.timestamp()), int(end_time.timestamp())
t_tmp_end_time = int(tmp_end_time.timestamp())
postdata['size'] = 10000
delta_time = 3600*24*3 # 7 days
sess = requests.Session()
sess.headers.update(headers)
def get_data(t, et):
if t >= et: return []
timestamp_range['gte'] = t*1000
timestamp_range['lt'] = et*1000
logging.info('Request data in [{}, {})'.format(datetime.datetime
|
ime, end_time))
|
conditional_block
|
query.py
|
备线路,1表示主线路
"over_drop", ##overlay非云网链路丢包率
"over_dropth", ##overlay非云网链路丢包率阈值
"over_sl_drop", ##overlay云网链路丢包率
"over_sl_dropth", ##overlay云网链路丢包率阈值
"over_delay", ##overlay链路时延
"over_shakedelay", ##overlay链路抖动
"over_mindelay",
"over_maxdelay",
"over_mdev",
"under_drop", ##underlay非云网链路丢包率
"under_dropth", ##underlay非云网链路丢包率阈值
"under_sl_drop", ##underlay云网链路丢包率
"under_sl_dropth", ##underlay云网链路丢包率阈值
"under_delay", ##underlay链路时延
"under_shakedelay", ##underlay链路抖动
"under_mindelay",
"under_maxdelay",
"under_mdev",
"tags",
#"metricset_name"
]
key2index = {k : i for i, k in enumerate(keys)}
def request_data(sess, postdata):
postdata = json.dumps(postdata)
response = sess.post(url, data=postdata)
res = response.text
data = json.loads(res)
samples = []
for i, a in enumerate(data['hits']['hits']):
pairs = a['_source']['pairs']
pair_dict = defaultdict(str)
pair_dict['timestamp'] = datetime.datetime.fromtimestamp(float(a['_source']['timestamp'])/1000)
if len(a['_source']['tags']) > 0:
pair_dict['tags'] = a['_source']['tags'][0]['value']
if len(a['_source']['tags']) > 1:
print('Warning:', a['_source']['tags'])
pair_dict['metricset_name'] = a['_source']['metricset']['name']
for p in pairs:
pair_dict[p['name']] = list(p['value'].values())[0]
samples.append([pair_dict[key] for key in keys])
return sorted(samples, key=lambda x:x[0]), data['hits']['total']
def get_list():
postdata = json.loads(json_query)
del postdata['query']['bool']['filter'][3:]
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
ta = datetime.datetime.now() - datetime.timedelta(hours=1)
tb = ta + datetime.timedelta(minutes=2)
timestamp_range['gte'] = ta.timestamp()*1000
timestamp_range['lt'] = tb.timestamp()*1000
logging.info('Request data in [{}, {})'.format(ta, tb))
sess = requests.Session()
sess.headers.update(headers)
for _ in range(3):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
break
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
df = pd.DataFrame(samples, columns=keys)
tunoip = {}
proname = {}
for nsid in df['ns_id'].unique():
if nsid=='': continue
t = df[df.ns_id==nsid]
tunoip[nsid] = sorted(list(t.tun_oip.unique()))
proname[nsid] = t.pop_proname.iloc[-1]
return tunoip, proname
def query(nsid, tunoip, begin_time=None, end_time=None, data_dir='./data'):
""" Query the ping data of specifical ns_id and tun_oip between specifical time range.
Time is datetime type or a string with format '%Y-%m-%d %H:%M:%S', e.g. '2019-10-13 10:09:00'
nsid: ns_id value.
tunoip: tun_oip value.
begin_time: The begining time to query. Default None means the time before 30 days.
end_time: The ending time to query. Default None means the current time.
"""
if type(begin_time) is str: begin_time = datetime.datetime.fromisoformat(begin_time)
if type(end_time) is str: end_time = datetime.datetime.fromisoformat(end_time)
nowtime = datetime.datetime.now()
pretime = nowtime - datetime.timedelta(days=30)
if end_time is None: end_time = nowtime
if begin_time is None: begin_time = pretime
q_begin_time, q_end_time = begin_time, end_time
if not os.path.exists(data_dir): os.mkdir(data_dir)
nsid_dir = os.path.join(data_dir, nsid)
if not os.path.exists(nsid_dir): os.mkdir(nsid_dir)
file_prefix = tunoip.replace('>', '')
# timestamp_path = os.path.join(nsid_dir, '%s.ts'%file_prefix)
csv_path = os.path.join(nsid_dir, '%s.csv'%file_prefix)
logging.info('Query: nsid={}, tunoip={}, timerange=[{}, {})'.format(nsid, tunoip, begin_time, end_time))
if os.path.exists(csv_path):
logging.info('Read data from %s'%csv_path)
df = pd.read_csv(csv_path)
logging.info('Time range: [{}, {}]'.format(df.iloc[0]['timestamp'], df.iloc[-1]['timestamp']))
begin_time = datetime.datetime.fromisoformat(df.iloc[-1]['timestamp']) + datetime.timedelta(minutes=1)
if begin_time < pretime:
begin_time = pretime
df.index = pd.to_datetime(df['timestamp'])
p_num = len(df)
if df.index[0] < pretime:
df = df.loc[pretime:]
logging.info('Forget %d records'%(p_num-len(df)))
else:
df = pd.DataFrame(columns=keys)
begin_time = pretime
# if nowtime - end_time < datetime.timedelta(hours=1):
# tmp_end_time = end_time - datetime.timedelta(hours=1)
# else:
# tmp_end_time = end_time
tmp_end_time = end_time
if q_end_time < begin_time:
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
return
s = json_query.replace('$nsid', nsid).replace('$tunoip', tunoip)
postdata = json.loads(s)
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
t_begin_time, t_end_time = int(begin_time.timestamp()), int(end_time.timestamp())
t_tmp_end_time = int(tmp_end_time.timestamp())
postdata['size'] = 10000
delta_time = 3600*24*3 # 7 days
sess = requests.Session()
sess.headers.update(headers)
def get_data(t, et):
if t >= et: return []
timestamp_range['gte'] = t*1000
timestamp_range['lt'] = et*1000
logging.info('Request data in [{}, {})'.format(datetime.datetime.fromtimestamp(t), datetime.datetime.fromtimestamp(et)))
for _ in range(5):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
return samples
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
data = []
for t in range(t_begin_time, t_tmp_end_time, delta_time):
et = min(t+delta_time, t_tmp_end_time)
data.extend(get_data(t, et))
df = df.append(pd.DataFrame(data, columns=keys), ignore_index=True)
# samples = get_data(t_tmp_end_time, t_end_time)
logging.info('Total Hists: {}'.format(len(data)))
if len(data)>0:
logging.info('Save data to %s'%csv_path)
df.to_csv(csv_path, index=False)
# df = df.append(pd.DataFrame(samples, columns=keys), ignore_index=True)
df.index = pd.to_datetime(df['timestamp'])
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
df.fillna(0, inplace=True)
return df
def cache():
tunoip, proname = get_list()
nsids = ['500850', '504608', '502902']
for nsid in nsids:
for lk in tunoip[nsid]:
query(nsid, lk, data_dir='/home/monitor/data')
# data_dir = '/home/monitor/data'
# fn = os.path.join(data_dir, 'collect.log')
fn = '/dev/stdou
|
t'
logging.basicConfig(level=logging.INFO,
format='[%(asctime)s] %(levelname)s: %(message)s',
datefmt='%Y-%m-%d %H:%M:%S')
if __name__ == "__main__":
|
identifier_body
|
|
query.py
|
filter": [
{"term": {"pairs.name": "tun_oip"}},
{"term": {"pairs.value.string": "$tunoip"}}
]
}
}
}
}
]
}
},
"size": 10000
}
'''
url = 'http://rpc.dsp.chinanetcenter.com:10200/api/console/proxy?path=*metricelf*%2F_search&method=POST'
headers = {
'Content-Type': 'application/json; charset=utf-8',
'Cookie': '__lcl=zh_CN; cluster=dashboard',
'DNT': '1',
'Host': 'rpc.dsp.chinanetcenter.com:10200',
'kbn-version': '6.1.1',
'Origin': 'http://rpc.dsp.chinanetcenter.com:10200',
'Referer': 'http://rpc.dsp.chinanetcenter.com:10200/app/kibana',
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/74.0.3729.108 Safari/537.36',
}
keys = [
"timestamp",
"pop_proname", ##项目名称
"pop_proid", ##oms上项目ID
"ns_id", ##项目的enterpriseID
"tun_uip", ##underlay链路ip
"tun_oip", ##overlay链路ip
"pop_ip", ##POP服务器管理ip
"tun_cost", ##链路权重
"tun_status", ##链路状态,0表示正常,1表示异常
"tun_cost_status", ##链路是否主备线路,0表示备线路,1表示主线路
"over_drop", ##overlay非云网链路丢包率
"over_dropth", ##overlay非云网链路丢包率阈值
"over_sl_drop", ##overlay云网链路丢包率
"over_sl_dropth", ##overlay云网链路丢包率阈值
"over_delay", ##overlay链路时延
"over_shakedelay", ##overlay链路抖动
"over_mindelay",
"over_maxdelay",
"over_mdev",
"under_drop", ##underlay非云网链路丢包率
"under_dropth", ##underlay非云网链路丢包率阈值
"under_sl_drop", ##underlay云网链路丢包率
"under_sl_dropth", ##underlay云网链路丢包率阈值
"under_delay", ##underlay链路时延
"under_shakedelay", ##underlay链路抖动
"under_mindelay",
"under_maxdelay",
"under_mdev",
"tags",
#"metricset_name"
]
key2index = {k : i for i, k in enumerate(keys)}
def request_data(sess, postdata):
postdata = json.dumps(postdata)
response = sess.post(url, data=postdata)
res = response.text
data = json.loads(res)
samples = []
for i, a in enumerate(data['hits']['hits']):
pairs = a['_source']['pairs']
|
defaultdict(str)
pair_dict['timestamp'] = datetime.datetime.fromtimestamp(float(a['_source']['timestamp'])/1000)
if len(a['_source']['tags']) > 0:
pair_dict['tags'] = a['_source']['tags'][0]['value']
if len(a['_source']['tags']) > 1:
print('Warning:', a['_source']['tags'])
pair_dict['metricset_name'] = a['_source']['metricset']['name']
for p in pairs:
pair_dict[p['name']] = list(p['value'].values())[0]
samples.append([pair_dict[key] for key in keys])
return sorted(samples, key=lambda x:x[0]), data['hits']['total']
def get_list():
postdata = json.loads(json_query)
del postdata['query']['bool']['filter'][3:]
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
ta = datetime.datetime.now() - datetime.timedelta(hours=1)
tb = ta + datetime.timedelta(minutes=2)
timestamp_range['gte'] = ta.timestamp()*1000
timestamp_range['lt'] = tb.timestamp()*1000
logging.info('Request data in [{}, {})'.format(ta, tb))
sess = requests.Session()
sess.headers.update(headers)
for _ in range(3):
try:
samples, tot = request_data(sess, postdata)
logging.info('Hists: {} / {}'.format(len(samples), tot))
break
except Exception as e:
logging.warning(e)
logging.info('Retry')
time.sleep(3)
df = pd.DataFrame(samples, columns=keys)
tunoip = {}
proname = {}
for nsid in df['ns_id'].unique():
if nsid=='': continue
t = df[df.ns_id==nsid]
tunoip[nsid] = sorted(list(t.tun_oip.unique()))
proname[nsid] = t.pop_proname.iloc[-1]
return tunoip, proname
def query(nsid, tunoip, begin_time=None, end_time=None, data_dir='./data'):
""" Query the ping data of specifical ns_id and tun_oip between specifical time range.
Time is datetime type or a string with format '%Y-%m-%d %H:%M:%S', e.g. '2019-10-13 10:09:00'
nsid: ns_id value.
tunoip: tun_oip value.
begin_time: The begining time to query. Default None means the time before 30 days.
end_time: The ending time to query. Default None means the current time.
"""
if type(begin_time) is str: begin_time = datetime.datetime.fromisoformat(begin_time)
if type(end_time) is str: end_time = datetime.datetime.fromisoformat(end_time)
nowtime = datetime.datetime.now()
pretime = nowtime - datetime.timedelta(days=30)
if end_time is None: end_time = nowtime
if begin_time is None: begin_time = pretime
q_begin_time, q_end_time = begin_time, end_time
if not os.path.exists(data_dir): os.mkdir(data_dir)
nsid_dir = os.path.join(data_dir, nsid)
if not os.path.exists(nsid_dir): os.mkdir(nsid_dir)
file_prefix = tunoip.replace('>', '')
# timestamp_path = os.path.join(nsid_dir, '%s.ts'%file_prefix)
csv_path = os.path.join(nsid_dir, '%s.csv'%file_prefix)
logging.info('Query: nsid={}, tunoip={}, timerange=[{}, {})'.format(nsid, tunoip, begin_time, end_time))
if os.path.exists(csv_path):
logging.info('Read data from %s'%csv_path)
df = pd.read_csv(csv_path)
logging.info('Time range: [{}, {}]'.format(df.iloc[0]['timestamp'], df.iloc[-1]['timestamp']))
begin_time = datetime.datetime.fromisoformat(df.iloc[-1]['timestamp']) + datetime.timedelta(minutes=1)
if begin_time < pretime:
begin_time = pretime
df.index = pd.to_datetime(df['timestamp'])
p_num = len(df)
if df.index[0] < pretime:
df = df.loc[pretime:]
logging.info('Forget %d records'%(p_num-len(df)))
else:
df = pd.DataFrame(columns=keys)
begin_time = pretime
# if nowtime - end_time < datetime.timedelta(hours=1):
# tmp_end_time = end_time - datetime.timedelta(hours=1)
# else:
# tmp_end_time = end_time
tmp_end_time = end_time
if q_end_time < begin_time:
df = df.loc[q_begin_time:q_end_time]
logging.info('Query: {} items'.format(len(df)))
logging.info('done.')
return
s = json_query.replace('$nsid', nsid).replace('$tunoip', tunoip)
postdata = json.loads(s)
timestamp_range = postdata['query']['bool']['filter'][2]['range']['timestamp']
t_begin_time, t_end_time = int(begin_time.timestamp()), int(end_time.timestamp())
t_tmp_end_time = int(tmp_end_time.timestamp())
postdata['size'] = 10000
delta_time = 3600*24*3 # 7 days
sess = requests.Session()
sess.headers.update(headers)
def get_data(t, et):
if t >= et: return []
timestamp_range['gte'] = t*1000
timestamp_range['lt'] = et*1000
logging.info('Request data in [{}, {})'.format(datetime.datetime
|
pair_dict =
|
identifier_name
|
day.page.ts
|
items = [];
total = [];
day = new Date().toString();
start = new Date().setHours(0, 0, 0, 0).toString();
end = new Date().setHours(23, 59, 59, 999).toString();
i = 0;
currency = '';
pdfObj = null;
company = '';
displaystart = '';
displayend = '';
a = 0;
salesvalue = 0;
min = '2020';
max = '';
constructor(
private router: Router,
private syncService: SyncService,
private loadingCtrl: LoadingController,
private datePipe: DatePipe,
private datepicker: DatePicker,
private plt: Platform,
private socialSharing: SocialSharing,
private file: File,
private fileOpener: FileOpener,
private storage: Storage,
private databaseservice: DatabaseService
)
|
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
this.start = first1;
this.end = last1;
// alert('start: ' + this.start + '\n' + this.end);
this.getData();
}
ngOnInit() {
}
share() {
// this.getData();
this.createPdf();
}
back() {
this.router.navigate(['/menu/reports']);
}
onActivate(event) {
if (event.type === 'click') {
console.log(event.row);
}
}
async getData() {
let loading = await this.loadingCtrl.create();
await loading.present();
let now = new Date().toString();
const a = this.datePipe.transform(now, 'dd MMM yyyy h:mm a');
let start = new Date().setHours(0, 0, 0, 0).toString();
const b = this.datePipe.transform(start, 'dd MMM yyyy h:mm a');
// alert('start:' + this.start + '\ncurrent:' + this.end);
// alert('day:' + this.day)
this.syncService.getTodaysSales(this.start, this.end, this.currency).then((data) => {
this.total = data;
console.log(data);
for (let i = 0; i < this.total.length; i++) {
const data1 = {
day: this.total[i].Date,
sales: this.total[i].Total,
paid: this.total[i].Paid
};
this.data.push(data1);
}
this.data = [...this.data];
loading.dismiss();
// alert('y' + this.total[0].Total + 'p' + this.total[0].Paid);
});
}
selectDate() {
var options={
date: new Date(),
mode: 'date',
androidTheme: this.datepicker.ANDROID_THEMES.THEME_DEVICE_DEFAULT_LIGHT
};
this.datepicker.show(options).then((date) => {
this.day = this.datePipe.transform(date, 'dd MMM yyyy h:mm a');
this.data = [];
this.getData();
// console.log('selected:',this.myDate);
});
}
changeDate() {
//alert('yes' + this.data.length);
//alert('y' + this.start);
if ( this.i === 2 || (this.a === 0 && this.i === this.salesvalue)) {
this.i = 0;
this.a = 1;
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
// const s = new Date(this.day).setHours(0, 0, 0, 0).toString();
// const l = new Date(this.day).setHours(23, 59, 59, 999).toString();
this.start = first1;
this.end = last1;
// this.start = this.datePipe.transform(s, 'dd MMM yyyy h:mm a');
// this.end = this.datePipe.transform(l, 'dd MMM yyyy h:mm a');
// alert('s: ' + first1 + '\nL: ' + last1 );
this.data = [];
this.getData();
}
this.i = this.i + 1;
this.day = this.datePipe.transform(this.day, 'dd MMM yyyy h:mm a');
}
async createPdf() {
let a ;
let x ;
let y ;
let z ;
let left;
let right;
let items = [];
let loading = await this.loadingCtrl.create();
await loading.present();
const itemss = [];
for (let i = 0; i < this.data.length; i++) {
itemss.push(
[
{ text: this.data[i].day.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].sales.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].paid.toString(), fontSize: 18, color: '#000000' }
]
);
}
left = {
text: [
{ text: this.company, bold: true, fontSize: 20, alignment: 'left'},
]
};
right = {
text: [
{ text: this.datePipe.transform(new Date(), 'dd MMM yyyy') , color: '#000000' , fontSize: 18, alignment: 'right'},
]
};
var docDefinition = {
pageSize: 'A4',
pageMargins: [ 20, 20, 20, 20 ],
content: [
{ text: 'SALES BY DAY', bold: true, alignment: 'center', fontSize: 25, style: 'subheader'},
{ text: this.displaystart + ' - ' + this.displayend, bold: true, alignment: 'center', fontSize: 20, style: 'subheader'},
// { margin: [0, 10, 0, 0],
// text: 'CUSTOMER STATEMENT', style: 'header', fontSize: 25, alignment: 'left', color: '#ff0000' },
{
margin: [0, 10, 0, 0],
canvas:
[
{
type: 'line',
x1: 0, y1: 0,
x2: 555, y2: 0,
lineWidth: 3
},
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [ 300, 250 ],
body: [
[
left, right
],
],
},
layout: 'noBorders'
},
{
margin: [0, 20, 0, 0],
text: [
{ text: 'Client ' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Total' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Paid' , style: 'subheader', bold: true
|
{
const last = new Date(new Date().getFullYear(), 11, 31);
this.max = this.datePipe.transform(last, 'yyyy');
this.storage.get('COM').then((val) => {
this.company = val;
});
storage.get('currency').then((val) => {
if (val !== null) {
this.currency = val.toString();
debugger
}
});
this.salesvalue = this.databaseservice.getSalesByDayValue();
this.day = this.datePipe.transform(this.day, 'MMM yyyy h:mm a');
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
|
identifier_body
|
day.page.ts
|
'\n' + this.end);
this.getData();
}
ngOnInit() {
}
share() {
// this.getData();
this.createPdf();
}
back() {
this.router.navigate(['/menu/reports']);
}
onActivate(event) {
if (event.type === 'click') {
console.log(event.row);
}
}
async getData() {
let loading = await this.loadingCtrl.create();
await loading.present();
let now = new Date().toString();
const a = this.datePipe.transform(now, 'dd MMM yyyy h:mm a');
let start = new Date().setHours(0, 0, 0, 0).toString();
const b = this.datePipe.transform(start, 'dd MMM yyyy h:mm a');
// alert('start:' + this.start + '\ncurrent:' + this.end);
// alert('day:' + this.day)
this.syncService.getTodaysSales(this.start, this.end, this.currency).then((data) => {
this.total = data;
console.log(data);
for (let i = 0; i < this.total.length; i++) {
const data1 = {
day: this.total[i].Date,
sales: this.total[i].Total,
paid: this.total[i].Paid
};
this.data.push(data1);
}
this.data = [...this.data];
loading.dismiss();
// alert('y' + this.total[0].Total + 'p' + this.total[0].Paid);
});
}
selectDate() {
var options={
date: new Date(),
mode: 'date',
androidTheme: this.datepicker.ANDROID_THEMES.THEME_DEVICE_DEFAULT_LIGHT
};
this.datepicker.show(options).then((date) => {
this.day = this.datePipe.transform(date, 'dd MMM yyyy h:mm a');
this.data = [];
this.getData();
// console.log('selected:',this.myDate);
});
}
changeDate() {
//alert('yes' + this.data.length);
//alert('y' + this.start);
if ( this.i === 2 || (this.a === 0 && this.i === this.salesvalue)) {
this.i = 0;
this.a = 1;
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
// const s = new Date(this.day).setHours(0, 0, 0, 0).toString();
// const l = new Date(this.day).setHours(23, 59, 59, 999).toString();
this.start = first1;
this.end = last1;
// this.start = this.datePipe.transform(s, 'dd MMM yyyy h:mm a');
// this.end = this.datePipe.transform(l, 'dd MMM yyyy h:mm a');
// alert('s: ' + first1 + '\nL: ' + last1 );
this.data = [];
this.getData();
}
this.i = this.i + 1;
this.day = this.datePipe.transform(this.day, 'dd MMM yyyy h:mm a');
}
async createPdf() {
let a ;
let x ;
let y ;
let z ;
let left;
let right;
let items = [];
let loading = await this.loadingCtrl.create();
await loading.present();
const itemss = [];
for (let i = 0; i < this.data.length; i++) {
itemss.push(
[
{ text: this.data[i].day.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].sales.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].paid.toString(), fontSize: 18, color: '#000000' }
]
);
}
left = {
text: [
{ text: this.company, bold: true, fontSize: 20, alignment: 'left'},
]
};
right = {
text: [
{ text: this.datePipe.transform(new Date(), 'dd MMM yyyy') , color: '#000000' , fontSize: 18, alignment: 'right'},
]
};
var docDefinition = {
pageSize: 'A4',
pageMargins: [ 20, 20, 20, 20 ],
content: [
{ text: 'SALES BY DAY', bold: true, alignment: 'center', fontSize: 25, style: 'subheader'},
{ text: this.displaystart + ' - ' + this.displayend, bold: true, alignment: 'center', fontSize: 20, style: 'subheader'},
// { margin: [0, 10, 0, 0],
// text: 'CUSTOMER STATEMENT', style: 'header', fontSize: 25, alignment: 'left', color: '#ff0000' },
{
margin: [0, 10, 0, 0],
canvas:
[
{
type: 'line',
x1: 0, y1: 0,
x2: 555, y2: 0,
lineWidth: 3
},
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [ 300, 250 ],
body: [
[
left, right
],
],
},
layout: 'noBorders'
},
{
margin: [0, 20, 0, 0],
text: [
{ text: 'Client ' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Total' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Paid' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [220, 160, 200 ],
body: itemss
},
layout: 'noBorders'
},
],
styles: {
header: {
fontSize: 18,
bold: true,
},
subheader: {
fontSize: 10,
bold: true,
// margin: [0, 15, 0, 0],
color: '#000000'
},
story: {
italic: true,
alignment: 'center',
width: '50%',
},
backgroundcolor: ''
}
};
this.pdfObj = pdfMake.createPdf(docDefinition);
this.downloadPdf(loading);
// });
}
ionViewWillLeave() {
this.databaseservice.setSalesByDayValue(2);
}
downloadPdf(loading) {
debugger
if (this.plt.is('cordova')) {
this.pdfObj.getBuffer((buffer) => {
// var blob = new Blob([buffer], { type: 'application/pdf' });
var utf8 = new Uint8Array(buffer);
var binaryArray = utf8.buffer;
var blob = new Blob([binaryArray], { type: 'application/pdf' });
const a = new Date().getTime().toString();
this.file.writeFile(this.file.dataDirectory, a + '.pdf', blob).then(fileEntry => {
// this.fileOpener.open(this.file.dataDirectory + a + '.pdf', 'application/pdf');
// loading.dismiss();
});
this.socialSharing.share("test", null, this.file.dataDirectory + a + '.pdf', null).then(() => {
loading.dismiss();
console.log('social share');
console.log(this.file.dataDirectory);
}).catch(() => {
});
debugger
// Save the PDF to the data Directory of our App
// this.file.writeFile(this.file.dataDirectory, 'Invoice4.pdf', blob).then(fileEntry => {
// this.fileOpener.open(this.file.dataDirectory + 'Invoice.pdf', 'application/pdf');
// });
});
} else
|
{
// On a browser simply use download!
this.pdfObj.download();
}
|
conditional_block
|
|
day.page.ts
|
, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
this.start = first1;
this.end = last1;
// alert('start: ' + this.start + '\n' + this.end);
this.getData();
}
ngOnInit() {
}
share() {
// this.getData();
this.createPdf();
}
back() {
this.router.navigate(['/menu/reports']);
}
onActivate(event) {
if (event.type === 'click') {
console.log(event.row);
}
}
async getData() {
let loading = await this.loadingCtrl.create();
await loading.present();
let now = new Date().toString();
const a = this.datePipe.transform(now, 'dd MMM yyyy h:mm a');
let start = new Date().setHours(0, 0, 0, 0).toString();
const b = this.datePipe.transform(start, 'dd MMM yyyy h:mm a');
// alert('start:' + this.start + '\ncurrent:' + this.end);
// alert('day:' + this.day)
this.syncService.getTodaysSales(this.start, this.end, this.currency).then((data) => {
this.total = data;
console.log(data);
for (let i = 0; i < this.total.length; i++) {
const data1 = {
day: this.total[i].Date,
sales: this.total[i].Total,
paid: this.total[i].Paid
};
this.data.push(data1);
}
this.data = [...this.data];
loading.dismiss();
// alert('y' + this.total[0].Total + 'p' + this.total[0].Paid);
});
}
selectDate() {
var options={
date: new Date(),
mode: 'date',
androidTheme: this.datepicker.ANDROID_THEMES.THEME_DEVICE_DEFAULT_LIGHT
};
this.datepicker.show(options).then((date) => {
this.day = this.datePipe.transform(date, 'dd MMM yyyy h:mm a');
this.data = [];
this.getData();
// console.log('selected:',this.myDate);
});
}
changeDate() {
//alert('yes' + this.data.length);
//alert('y' + this.start);
if ( this.i === 2 || (this.a === 0 && this.i === this.salesvalue)) {
this.i = 0;
this.a = 1;
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
// const s = new Date(this.day).setHours(0, 0, 0, 0).toString();
// const l = new Date(this.day).setHours(23, 59, 59, 999).toString();
this.start = first1;
this.end = last1;
// this.start = this.datePipe.transform(s, 'dd MMM yyyy h:mm a');
// this.end = this.datePipe.transform(l, 'dd MMM yyyy h:mm a');
// alert('s: ' + first1 + '\nL: ' + last1 );
this.data = [];
this.getData();
}
this.i = this.i + 1;
this.day = this.datePipe.transform(this.day, 'dd MMM yyyy h:mm a');
}
async createPdf() {
let a ;
let x ;
let y ;
let z ;
let left;
let right;
let items = [];
let loading = await this.loadingCtrl.create();
await loading.present();
const itemss = [];
for (let i = 0; i < this.data.length; i++) {
itemss.push(
[
{ text: this.data[i].day.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].sales.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].paid.toString(), fontSize: 18, color: '#000000' }
]
);
}
left = {
text: [
{ text: this.company, bold: true, fontSize: 20, alignment: 'left'},
]
};
right = {
text: [
{ text: this.datePipe.transform(new Date(), 'dd MMM yyyy') , color: '#000000' , fontSize: 18, alignment: 'right'},
]
};
var docDefinition = {
pageSize: 'A4',
pageMargins: [ 20, 20, 20, 20 ],
content: [
{ text: 'SALES BY DAY', bold: true, alignment: 'center', fontSize: 25, style: 'subheader'},
{ text: this.displaystart + ' - ' + this.displayend, bold: true, alignment: 'center', fontSize: 20, style: 'subheader'},
// { margin: [0, 10, 0, 0],
// text: 'CUSTOMER STATEMENT', style: 'header', fontSize: 25, alignment: 'left', color: '#ff0000' },
{
margin: [0, 10, 0, 0],
canvas:
[
{
type: 'line',
x1: 0, y1: 0,
x2: 555, y2: 0,
lineWidth: 3
},
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [ 300, 250 ],
body: [
[
left, right
],
],
},
layout: 'noBorders'
},
{
margin: [0, 20, 0, 0],
text: [
{ text: 'Client ' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Total' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Paid' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [220, 160, 200 ],
body: itemss
},
layout: 'noBorders'
},
],
styles: {
header: {
fontSize: 18,
bold: true,
},
subheader: {
fontSize: 10,
bold: true,
// margin: [0, 15, 0, 0],
color: '#000000'
},
story: {
italic: true,
alignment: 'center',
width: '50%',
},
backgroundcolor: ''
}
};
this.pdfObj = pdfMake.createPdf(docDefinition);
this.downloadPdf(loading);
// });
}
ionViewWillLeave() {
this.databaseservice.setSalesByDayValue(2);
}
downloadPdf(loading) {
debugger
if (this.plt.is('cordova')) {
this.pdfObj.getBuffer((buffer) => {
// var blob = new Blob([buffer], { type: 'application/pdf' });
var utf8 = new Uint8Array(buffer);
var binaryArray = utf8.buffer;
var blob = new Blob([binaryArray], { type: 'application/pdf' });
const a = new Date().getTime().toString();
this.file.writeFile(this.file.dataDirectory, a + '.pdf', blob).then(fileEntry => {
// this.fileOpener.open(this.file.dataDirectory + a + '.pdf', 'application/pdf');
// loading.dismiss();
});
this.socialSharing.share("test", null, this.file.dataDirectory + a + '.pdf', null).then(() => {
loading.dismiss();
console.log('social share');
console.log(this.file.dataDirectory);
}).catch(() => {
|
});
debugger
// Save the PDF to the data Directory of our App
|
random_line_split
|
|
day.page.ts
|
implements OnInit {
style = 'bootstrap';
data = [];
items = [];
total = [];
day = new Date().toString();
start = new Date().setHours(0, 0, 0, 0).toString();
end = new Date().setHours(23, 59, 59, 999).toString();
i = 0;
currency = '';
pdfObj = null;
company = '';
displaystart = '';
displayend = '';
a = 0;
salesvalue = 0;
min = '2020';
max = '';
constructor(
private router: Router,
private syncService: SyncService,
private loadingCtrl: LoadingController,
private datePipe: DatePipe,
private datepicker: DatePicker,
private plt: Platform,
private socialSharing: SocialSharing,
private file: File,
private fileOpener: FileOpener,
private storage: Storage,
private databaseservice: DatabaseService
) {
const last = new Date(new Date().getFullYear(), 11, 31);
this.max = this.datePipe.transform(last, 'yyyy');
this.storage.get('COM').then((val) => {
this.company = val;
});
storage.get('currency').then((val) => {
if (val !== null) {
this.currency = val.toString();
debugger
}
});
this.salesvalue = this.databaseservice.getSalesByDayValue();
this.day = this.datePipe.transform(this.day, 'MMM yyyy h:mm a');
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
this.start = first1;
this.end = last1;
// alert('start: ' + this.start + '\n' + this.end);
this.getData();
}
ngOnInit() {
}
share() {
// this.getData();
this.createPdf();
}
back() {
this.router.navigate(['/menu/reports']);
}
onActivate(event) {
if (event.type === 'click') {
console.log(event.row);
}
}
async getData() {
let loading = await this.loadingCtrl.create();
await loading.present();
let now = new Date().toString();
const a = this.datePipe.transform(now, 'dd MMM yyyy h:mm a');
let start = new Date().setHours(0, 0, 0, 0).toString();
const b = this.datePipe.transform(start, 'dd MMM yyyy h:mm a');
// alert('start:' + this.start + '\ncurrent:' + this.end);
// alert('day:' + this.day)
this.syncService.getTodaysSales(this.start, this.end, this.currency).then((data) => {
this.total = data;
console.log(data);
for (let i = 0; i < this.total.length; i++) {
const data1 = {
day: this.total[i].Date,
sales: this.total[i].Total,
paid: this.total[i].Paid
};
this.data.push(data1);
}
this.data = [...this.data];
loading.dismiss();
// alert('y' + this.total[0].Total + 'p' + this.total[0].Paid);
});
}
selectDate() {
var options={
date: new Date(),
mode: 'date',
androidTheme: this.datepicker.ANDROID_THEMES.THEME_DEVICE_DEFAULT_LIGHT
};
this.datepicker.show(options).then((date) => {
this.day = this.datePipe.transform(date, 'dd MMM yyyy h:mm a');
this.data = [];
this.getData();
// console.log('selected:',this.myDate);
});
}
changeDate() {
//alert('yes' + this.data.length);
//alert('y' + this.start);
if ( this.i === 2 || (this.a === 0 && this.i === this.salesvalue)) {
this.i = 0;
this.a = 1;
const firstdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth(), 1).setHours(0, 0, 0, 0);
const first1 = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy h:mm a');
const lastdayofmonth = new Date(new Date(this.day).getFullYear(), new Date(this.day).getMonth() + 1, 0).setHours(23, 59, 59, 999);
const last1 = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy h:mm a');
this.displaystart = this.datePipe.transform(firstdayofmonth, 'dd MMM yyyy');
this.displayend = this.datePipe.transform(lastdayofmonth, 'dd MMM yyyy');
// const s = new Date(this.day).setHours(0, 0, 0, 0).toString();
// const l = new Date(this.day).setHours(23, 59, 59, 999).toString();
this.start = first1;
this.end = last1;
// this.start = this.datePipe.transform(s, 'dd MMM yyyy h:mm a');
// this.end = this.datePipe.transform(l, 'dd MMM yyyy h:mm a');
// alert('s: ' + first1 + '\nL: ' + last1 );
this.data = [];
this.getData();
}
this.i = this.i + 1;
this.day = this.datePipe.transform(this.day, 'dd MMM yyyy h:mm a');
}
async createPdf() {
let a ;
let x ;
let y ;
let z ;
let left;
let right;
let items = [];
let loading = await this.loadingCtrl.create();
await loading.present();
const itemss = [];
for (let i = 0; i < this.data.length; i++) {
itemss.push(
[
{ text: this.data[i].day.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].sales.toString() , fontSize: 18, color: '#000000' },
{ text: this.data[i].paid.toString(), fontSize: 18, color: '#000000' }
]
);
}
left = {
text: [
{ text: this.company, bold: true, fontSize: 20, alignment: 'left'},
]
};
right = {
text: [
{ text: this.datePipe.transform(new Date(), 'dd MMM yyyy') , color: '#000000' , fontSize: 18, alignment: 'right'},
]
};
var docDefinition = {
pageSize: 'A4',
pageMargins: [ 20, 20, 20, 20 ],
content: [
{ text: 'SALES BY DAY', bold: true, alignment: 'center', fontSize: 25, style: 'subheader'},
{ text: this.displaystart + ' - ' + this.displayend, bold: true, alignment: 'center', fontSize: 20, style: 'subheader'},
// { margin: [0, 10, 0, 0],
// text: 'CUSTOMER STATEMENT', style: 'header', fontSize: 25, alignment: 'left', color: '#ff0000' },
{
margin: [0, 10, 0, 0],
canvas:
[
{
type: 'line',
x1: 0, y1: 0,
x2: 555, y2: 0,
lineWidth: 3
},
]
},
{
margin: [0, 5, 0, 0],
style: 'totalsTable',
table: {
widths: [ 300, 250 ],
body: [
[
left, right
],
],
},
layout: 'noBorders'
},
{
margin: [0, 20, 0, 0],
text: [
{ text: 'Client ' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text: ' Total' , style: 'subheader', bold: true, fontSize: 20, color: '#000000' },
{ text:
|
DayPage
|
identifier_name
|
|
models.py
|
def forward(self, x):
return self.act(self.bn(self.conv(x)))
def fuse_forward(self, x):
# 合并后的前向推理,bn和卷积合并
return self.act(self.conv(x))
class Bottleneck(nn.Module):
'''
标准瓶颈层
'''
def __init__(self, in_channel, out_channel, shortcut=True, groups=1, expansion=0.5): # ch_in, ch_out, shortcut, groups, expansion
super(Bottleneck, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel, out_channel, 3, 1, groups=groups)
self.add = shortcut and in_channel == out_channel
def forward(self, x):
return x + self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))
class BottleneckCSP(nn.Module):
# CSP Bottleneck https://github.com/WongKinYiu/CrossStagePartialNetworks
def __init__(self, in_channel, out_channel, repeats=1, shortcut=True, groups=1, expansion=0.5):
super(BottleneckCSP, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = nn.Conv2d(in_channel, hidden_channel, 1, 1, bias=False)
self.cv3 = nn.Conv2d(hidden_channel, hidden_channel, 1, 1, bias=False)
self.cv4 = Conv(2 * hidden_channel, out_channel, 1, 1)
self.bn = nn.BatchNorm2d(2 * hidden_channel) # applied to cat(cv2, cv3)
self.act = nn.LeakyReLU(0.1, inplace=True)
self.m = nn.Sequential(*[Bottleneck(hidden_channel, hidden_channel, shortcut, groups, expansion=1.0) for _ in range(repeats)])
def forward(self, x):
y1 = self.cv3(self.m(self.cv1(x)))
y2 = self.cv2(x)
return self.cv4(self.act(self.bn(torch.cat((y1, y2), dim=1))))
class SPP(nn.Module):
# Spatial pyramid pooling layer used in YOLOv3-SPP
def __init__(self, in_channel, out_channel, kernel_size_list=(5, 9, 13)):
super(SPP, self).__init__()
hidden_channel = in_channel // 2 # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel * (len(kernel_size_list) + 1), out_channel, 1, 1)
self.m = nn.ModuleList([nn.MaxPool2d(kernel_size=kernel_size, stride=1, padding=kernel_size // 2) for kernel_size in kernel_size_list])
def forward(self, x):
x = self.cv1(x)
return self.cv2(torch.cat([x] + [m(x) for m in self.m], dim=1))
class Focus(nn.Module):
# Focus wh information into c-space
def __init__(self, in_channel, out_channel, kernel_size=1, stride=1, padding=None, groups=1, activation=True):
super(Focus, self).__init__()
self.conv = Conv(in_channel * 4, out_channel, kernel_size, stride, padding, groups, activation)
def forward(self, x): # x(b,c,w,h) -> y(b,4c,w/2,h/2)
return self.conv(torch.cat([x[..., ::2, ::2], x[..., 1::2, ::2], x[..., ::2, 1::2], x[..., 1::2, 1::2]], dim=1))
class Concat(nn.Module):
# Concatenate a list of tensors along dimension
def __init__(self, dimension=1):
super(Concat, self).__init__()
self.d = dimension
def forward(self, x):
return torch.cat(x, dim=self.d)
class Detect(nn.Module):
def __init__(self, num_classes, num_anchor, reference_channels):
super(Detect, self).__init__()
self.num_anchor = num_anchor
self.num_classes = num_classes
self.num_output = self.num_classes + 5
self.m = nn.ModuleList(nn.Conv2d(input_channel, self.num_output * self.num_anchor, 1) for input_channel in reference_channels)
self.init_weight()
def forward(self, x):
for ilevel, module in enumerate(self.m):
x[ilevel] = module(x[ilevel])
return x
def init_weight(self):
strides = [8, 16, 32]
for head, stride in zip(self.m, strides):
bias = head.bias.view(self.num_anchor, -1)
bias[:, 4] += math.log(8 / (640 / stride) ** 2)
bias[:, 5:] += math.log(0.6 / (self.num_classes - 0.99))
head.bias = nn.Parameter(bias.view(-1), requires_grad=True)
class Yolo(nn.Module):
def __init__(self, num_classes, config_file, rank=0):
super(Yolo, self).__init__()
self.num_classes = num_classes
self.rank = rank
self.strides = [8, 16, 32]
self.model, self.saved_index, anchors = self.build_model(config_file)
self.register_buffer("anchors", torch.FloatTensor(anchors).view(3, 3, 2) / torch.FloatTensor(self.strides).view(3, 1, 1))
self.apply(self.init_weight)
def set_new_anchors(self, anchors):
# 对设置的anchors缩放到特征图大小
self.anchors[...] = anchors / torch.FloatTensor(self.strides).view(3, 1, 1)
def init_weight(self, m):
type_t = type(m)
if type_t is nn.Conv2d:
# pass init
pass
elif type_t is nn.BatchNorm2d:
m.eps = 1e-3
m.momentum = 0.03
elif type_t in [nn.LeakyReLU, nn.ReLU, nn.ReLU6]:
m.inplace = True
def forward(self, x):
y = []
for module in self.model:
if module.from_index != -1:
if isinstance(module.from_index, int):
x = y[module.from_index]
else:
xout = []
for i in module.from_index:
if i == -1:
xval = x
else:
xval = y[i]
xout.append(xval)
x = xout
x = module(x)
y.append(x if module.layer_index in self.saved_index else None)
return x
def parse_string(self, value):
if value == "None":
return None
elif value == "True":
return True
elif value == "False":
return False
else:
return value
def fuse(self): # fuse model Conv2d() + BatchNorm2d() layers
print('Fusing layers... ', end='')
for m in self.model.modules():
if type(m) is Conv:
m.conv = nn_utils.fuse_conv_and_bn(m.conv, m.bn) # update conv
m.bn = None # remove batchnorm
m.forward = m.fuse_forward # update forward
return self
def build_model(self, config_file, input_channel=3):
with open(config_file) as f:
self.yaml = yaml.load(f, Loader=yaml.FullLoader)
all_layers_cfg_list = self.yaml["backbone"] + self.yaml["head"]
anchors, depth_multiple, width_multiple = [self.yaml[item] for item in ["anchors", "depth_multiple", "width_multiple"]]
num_classes = self.num_classes
num_anchor = len(anchors[0]) // 2
num_output = num_anchor * (num_classes + 5)
|
args = [self.parse_string(a) for a in args]
module_function = eval(module_name)
if repeat_count > 1:
repeat_count = max(round(repeat_count * depth_multiple), 1)
if module_function in [Conv, Bottleneck, SPP, Focus, BottleneckCSP]:
channel_input, channel_output = all_layers_channels[from_index], args[0]
if channel_output != num_output:
channel_output = make_divisible(channel_output * width_multiple, 8)
args = [channel_input, channel_output, *args[1:]]
if module_function in [BottleneckCSP]:
args.insert(2, repeat_count)
repeat_count = 1
elif module_function is Concat:
channel_output = sum([all_layers_channels[-1 if x == -1 else x + 1] for x in from_index])
elif
|
all_layers_channels = [input_channel]
all_layers = []
saved_layer_index = []
for layer_index, (from_index, repeat_count, module_name, args) in enumerate(all_layers_cfg_list):
|
random_line_split
|
models.py
|
def autopad(kernel, padding=None): # kernel, padding
# Pad to 'same'
if padding is None:
padding = kernel // 2 if isinstance(kernel, int) else [x // 2 for x in kernel] # auto-pad
return padding
class Conv(nn.Module):
'''
标准卷积层
'''
def __init__(self, in_channel, out_channel, kernel_size=1, stride=1, padding=None, groups=1, activation=True):
super(Conv, self).__init__()
self.conv = nn.Conv2d(in_channel, out_channel, kernel_size, stride, autopad(kernel_size, padding), groups=groups, bias=False)
self.bn = nn.BatchNorm2d(out_channel)
self.act = nn.LeakyReLU(0.1, inplace=True) if activation else nn.Identity()
def forward(self, x):
return self.act(self.bn(self.conv(x)))
def fuse_forward(self, x):
# 合并后的前向推理,bn和卷积合并
return self.act(self.conv(x))
class Bottleneck(nn.Module):
'''
标准瓶颈层
'''
def __init__(self, in_channel, out_channel, shortcut=True, groups=1, expansion=0.5): # ch_in, ch_out, shortcut, groups, expansion
super(Bottleneck, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel, out_channel, 3, 1, groups=groups)
self.add = shortcut and in_channel == out_channel
def forward(self, x):
return x + self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))
class BottleneckCSP(nn.Module):
# CSP Bottleneck https://github.com/WongKinYiu/CrossStagePartialNetworks
def __init__(self, in_channel, out_channel, repeats=1, shortcut=True, groups=1, expansion=0.5):
super(BottleneckCSP, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = nn.Conv2d(in_channel, hidden_channel, 1, 1, bias=False)
self.cv3 = nn.Conv2d(hidden_channel, hidden_channel, 1, 1, bias=False)
self.cv4 = Conv(2 * hidden_channel, out_channel, 1, 1)
self.bn = nn.BatchNorm2d(2 * hidden_channel) # applied to cat(cv2, cv3)
self.act = nn.LeakyReLU(0.1, inplace=True)
self.m = nn.Sequential(*[Bottleneck(hidden_channel, hidden_channel, shortcut, groups, expansion=1.0) for _ in range(repeats)])
def forward(self, x):
y1 = self.cv3(self.m(self.cv1(x)))
y2 = self.cv2(x)
return self.cv4(self.act(self.bn(torch.cat((y1, y2), dim=1))))
class SPP(nn.Module):
# Spatial pyramid pooling layer used in YOLOv3-SPP
def __init__(self, in_channel, out_channel, kernel_size_list=(5, 9, 13)):
super(SPP, self).__init__()
hidden_channel = in_channel // 2 # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel * (len(kernel_size_list) + 1), out_channel, 1, 1)
self.m = nn.ModuleList([nn.MaxPool2d(kernel_size=kernel_size, stride=1, padding=kernel_size // 2) for kernel_size in kernel_size_list])
def forward(self, x):
x = self.cv1(x)
return self.cv2(torch.cat([x] + [m(x) for m in self.m], dim=1))
class Focus(nn.Module):
# Focus wh information into c-space
def __init__(self, in_channel, out_channel, kernel_size=1, stride=1, padding=None, groups=1, activation=True):
super(Focus, self).__init__()
self.conv = Conv(in_channel * 4, out_channel, kernel_size, stride, padding, groups, activation)
def forward(self, x): # x(b,c,w,h) -> y(b,4c,w/2,h/2)
return self.conv(torch.cat([x[..., ::2, ::2], x[..., 1::2, ::2], x[..., ::2, 1::2], x[..., 1::2, 1::2]], dim=1))
class Concat(nn.Module):
# Concatenate a list of tensors along dimension
def __init__(self, dimension=1):
super(Concat, self).__init__()
self.d = dimension
def forward(self, x):
return torch.cat(x, dim=self.d)
class Detect(nn.Module):
def __init__(self, num_classes, num_anchor, reference_channels):
super(Detect, self).__init__()
self.num_anchor = num_anchor
self.num_classes = num_classes
self.num_output = self.num_classes + 5
self.m = nn.ModuleList(nn.Conv2d(input_channel, self.num_output * self.num_anchor, 1) for input_channel in reference_channels)
self.init_weight()
def forward(self, x):
for ilevel, module in enumerate(self.m):
x[ilevel] = module(x[ilevel])
return x
def init_weight(self):
strides = [8, 16, 32]
for head, stride in zip(self.m, strides):
bias = head.bias.view(self.num_anchor, -1)
bias[:, 4] += math.log(8 / (640 / stride) ** 2)
bias[:, 5:] += math.log(0.6 / (self.num_classes - 0.99))
head.bias = nn.Parameter(bias.view(-1), requires_grad=True)
class Yolo(nn.Module):
def __init__(self, num_classes, config_file, rank=0):
super(Yolo, self).__init__()
self.num_classes = num_classes
self.rank = rank
self.strides = [8, 16, 32]
self.model, self.saved_index, anchors = self.build_model(config_file)
self.register_buffer("anchors", torch.FloatTensor(anchors).view(3, 3, 2) / torch.FloatTensor(self.strides).view(3, 1, 1))
self.apply(self.init_weight)
def set_new_anchors(self, anchors):
# 对设置的anchors缩放到特征图大小
self.anchors[...] = anchors / torch.FloatTensor(self.strides).view(3, 1, 1)
def init_weight(self, m):
type_t = type(m)
if type_t is nn.Conv2d:
# pass init
pass
elif type_t is nn.BatchNorm2d:
m.eps = 1e-3
m.momentum = 0.03
elif type_t in [nn.LeakyReLU, nn.ReLU, nn.ReLU6]:
m.inplace = True
def forward(self, x):
y = []
for module in self.model:
if module.from_index != -1:
if isinstance(module.from_index, int):
x = y[module.from_index]
else:
xout = []
for i in module.from_index:
if i == -1:
xval = x
else:
xval = y[i]
xout.append(xval)
x = xout
x = module(x)
y.append(x if module.layer_index in self.saved_index else None)
return x
def parse_string(self, value):
if value == "None":
return None
elif value == "True":
return True
elif value == "False":
return False
else:
return value
def fuse(self): # fuse model Conv2d() + BatchNorm2d() layers
print('Fusing layers... ', end='')
for m in self.model.modules():
if type(m) is Conv:
m.conv = nn_utils.fuse_conv_and_bn(m.conv, m.bn) # update conv
m.bn = None # remove batchnorm
m.forward = m.fuse_forward # update forward
return self
def build_model(self, config_file, input_channel=3):
with open(config_file) as f:
self.yaml = yaml.load(f, Loader=yaml.FullLoader)
all_layers_cfg_list = self.yaml["backbone"] + self.yaml["head"]
anchors, depth_multiple, width_multiple = [self.yaml[item] for item in ["anchors", "depth_multiple", "width_multiple"]]
num_classes = self.num_classes
num_anchor = len(anchors[0]) // 2
num_output = num_anchor * (num_classes + 5)
all_layers_channels = [input_channel]
all_layers = []
saved_layer_index = []
for layer_index, (from_index, repeat_count, module_name, args) in enumerate(all
|
return math.ceil(x / divisor) * divisor
|
identifier_body
|
|
models.py
|
forward(self, x):
return self.act(self.bn(self.conv(x)))
def fuse_forward(self, x):
# 合并后的前向推理,bn和卷积合并
return self.act(self.conv(x))
class Bottleneck(nn.Module):
'''
标准瓶颈层
'''
def __init__(self, in_channel, out_channel, shortcut=True, groups=1, expansion=0.5): # ch_in, ch_out, shortcut, groups, expansion
super(Bottleneck, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel, out_channel, 3, 1, groups=groups)
self.add = shortcut and in_channel == out_channel
def forward(self, x):
return x + self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))
class BottleneckCSP(nn.Module):
# CSP Bottleneck https://github.com/WongKinYiu/CrossStagePartialNetworks
def __init__(self, in_channel, out_channel, repeats=1, shortcut=True, groups=1, expansion=0.5):
super(BottleneckCSP, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = nn.Conv2d(in_channel, hidden_channel, 1, 1, bias=False)
self.cv3 = nn.Conv2d(hidden_channel, hidden_channel, 1, 1, bias=False)
self.cv4 = Conv(2 * hidden_channel, out_channel, 1, 1)
self.bn = nn.BatchNorm2d(2 * hidden_channel) # applied to cat(cv2, cv3)
self.act = nn.LeakyReLU(0.1, inplace=True)
self.m = nn.Sequential(*[Bottleneck(hidden_channel, hidden_channel, shortcut, groups, expansion=1.0) for _ in range(repeats)])
def forward(self, x):
y1 = self.cv3(self.m(self.cv1(x)))
y2 = self.cv2(x)
return self.cv4(self.act(self.bn(torch.cat((y1, y2), dim=1))))
class SPP(nn.Module):
# Spatial pyramid pooling layer used in YOLOv3-SPP
def __init__(self, in_channel, out_channel, kernel_size_list=(5, 9, 13)):
super(SPP, self).__init__()
hidden_channel = in_channel // 2 # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel * (len(kernel_size_list) + 1), out_channel, 1, 1)
self.m = nn.ModuleList([nn.MaxPool2d(kernel_size=kernel_size, stride=1, padding=kernel_size // 2) for kernel_size in kernel_size_list])
def forward(self, x):
x = self.cv1(x)
return self.cv2(torch.cat([x] + [m(x) for m in self.m], dim=1))
class Focus(nn.Module):
# Focus wh information into c-space
def __init__(self, in_channel, out_channel, kernel_size=1, stride=1, padding=None, groups=1, activation=True):
super(Focus, self).__init__()
self.conv = Conv(in_channel * 4, out_channel, kernel_size, stride, padding, groups, activation)
def forward(self, x): # x(b,c,w,h) -> y(b,4c,w/2,h/2)
return self.conv(torch.cat([x[..., ::2, ::2], x[..., 1::2, ::2], x[..., ::2, 1::2], x[..., 1::2, 1::2]], dim=1))
class Concat(nn.Module):
# Concatenate a list of tensors along dimension
def __init__(self, dimension=1):
super(Concat, self).__init__()
self.d = dimension
def forward(self, x):
return torch.cat(x, dim=self.d)
class Detect(nn.Module):
def __init__(self, num_classes, num_anchor, reference_channels):
super(Detect, self).__init__()
self.num_anchor = num_anchor
self.num_classes = num_classes
self.num_output = self.num_classes + 5
self.m = nn.ModuleList(nn.Conv2d(input_channel, self.num_output * self.num_anchor, 1) for input_channel in reference_channels)
self.init_weight()
def forward(self, x):
for ilevel, module in enumerate(self.m):
x[ilevel] = module(x[ilevel])
return x
|
strides = [8, 16, 32]
for head, stride in zip(self.m, strides):
bias = head.bias.view(self.num_anchor, -1)
bias[:, 4] += math.log(8 / (640 / stride) ** 2)
bias[:, 5:] += math.log(0.6 / (self.num_classes - 0.99))
head.bias = nn.Parameter(bias.view(-1), requires_grad=True)
class Yolo(nn.Module):
def __init__(self, num_classes, config_file, rank=0):
super(Yolo, self).__init__()
self.num_classes = num_classes
self.rank = rank
self.strides = [8, 16, 32]
self.model, self.saved_index, anchors = self.build_model(config_file)
self.register_buffer("anchors", torch.FloatTensor(anchors).view(3, 3, 2) / torch.FloatTensor(self.strides).view(3, 1, 1))
self.apply(self.init_weight)
def set_new_anchors(self, anchors):
# 对设置的anchors缩放到特征图大小
self.anchors[...] = anchors / torch.FloatTensor(self.strides).view(3, 1, 1)
def init_weight(self, m):
type_t = type(m)
if type_t is nn.Conv2d:
# pass init
pass
elif type_t is nn.BatchNorm2d:
m.eps = 1e-3
m.momentum = 0.03
elif type_t in [nn.LeakyReLU, nn.ReLU, nn.ReLU6]:
m.inplace = True
def forward(self, x):
y = []
for module in self.model:
if module.from_index != -1:
if isinstance(module.from_index, int):
x = y[module.from_index]
else:
xout = []
for i in module.from_index:
if i == -1:
xval = x
else:
xval = y[i]
xout.append(xval)
x = xout
x = module(x)
y.append(x if module.layer_index in self.saved_index else None)
return x
def parse_string(self, value):
if value == "None":
return None
elif value == "True":
return True
elif value == "False":
return False
else:
return value
def fuse(self): # fuse model Conv2d() + BatchNorm2d() layers
print('Fusing layers... ', end='')
for m in self.model.modules():
if type(m) is Conv:
m.conv = nn_utils.fuse_conv_and_bn(m.conv, m.bn) # update conv
m.bn = None # remove batchnorm
m.forward = m.fuse_forward # update forward
return self
def build_model(self, config_file, input_channel=3):
with open(config_file) as f:
self.yaml = yaml.load(f, Loader=yaml.FullLoader)
all_layers_cfg_list = self.yaml["backbone"] + self.yaml["head"]
anchors, depth_multiple, width_multiple = [self.yaml[item] for item in ["anchors", "depth_multiple", "width_multiple"]]
num_classes = self.num_classes
num_anchor = len(anchors[0]) // 2
num_output = num_anchor * (num_classes + 5)
all_layers_channels = [input_channel]
all_layers = []
saved_layer_index = []
for layer_index, (from_index, repeat_count, module_name, args) in enumerate(all_layers_cfg_list):
args = [self.parse_string(a) for a in args]
module_function = eval(module_name)
if repeat_count > 1:
repeat_count = max(round(repeat_count * depth_multiple), 1)
if module_function in [Conv, Bottleneck, SPP, Focus, BottleneckCSP]:
channel_input, channel_output = all_layers_channels[from_index], args[0]
if channel_output != num_output:
channel_output = make_divisible(channel_output * width_multiple, 8)
args = [channel_input, channel_output, *args[1:]]
if module_function in [BottleneckCSP]:
args.insert(2, repeat_count)
repeat_count = 1
elif module_function is Concat:
channel_output = sum([all_layers_channels[-1 if x == -1 else x + 1] for x in from_index])
elif
|
def init_weight(self):
|
conditional_block
|
models.py
|
forward(self, x):
return self.act(self.bn(self.conv(x)))
def fuse_forward(self, x):
# 合并后的前向推理,bn和卷积合并
return self.act(self.conv(x))
class Bottleneck(nn.Module):
'''
标准瓶颈层
'''
def __init__(self, in_channel, out_channel, shortcut=True, groups=1, expansion=0.5): # ch_in, ch_out, shortcut, groups, expansion
super(Bottleneck, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel, out_channel, 3, 1, groups=groups)
self.add = shortcut and in_channel == out_channel
def forward(self, x):
return x + self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))
class BottleneckCSP(nn.Module):
# CSP Bottleneck https://github.com/WongKinYiu/CrossStagePartialNetworks
def __init__(self, in_channel, out_channel, repeats=1, shortcut=True, groups=1, expansion=0.5):
super(BottleneckCSP, self).__init__()
hidden_channel = int(out_channel * expansion) # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = nn.Conv2d(in_channel, hidden_channel, 1, 1, bias=False)
self.cv3 = nn.Conv2d(hidden_channel, hidden_channel, 1, 1, bias=False)
self.cv4 = Conv(2 * hidden_channel, out_channel, 1, 1)
self.bn = nn.BatchNorm2d(2 * hidden_channel) # applied to cat(cv2, cv3)
self.act = nn.LeakyReLU(0.1, inplace=True)
self.m = nn.Sequential(*[Bottleneck(hidden_channel, hidden_channel, shortcut, groups, expansion=1.0) for _ in range(repeats)])
def forward(self, x):
y1 = self.cv3(self.m(self.cv1(x)))
y2 = self.cv2(x)
return self.cv4(self.act(self.bn(torch.cat((y1, y2), dim=1))))
class SPP(nn.Module):
# Spatial pyramid pooling layer used in YOLOv3-SPP
def __init__(self, in_channel, out_channel, kernel_size_list=(5, 9, 13)):
super(SPP, self).__init__()
hidden_channel = in_channel // 2 # hidden channels
self.cv1 = Conv(in_channel, hidden_channel, 1, 1)
self.cv2 = Conv(hidden_channel * (len(kernel_size_list) + 1), out_channel, 1, 1)
self.m = nn.ModuleList([nn.MaxPool2d(kernel_size=kernel_size, stride=1, padding=kernel_size // 2) for kernel_size in kernel_size_list])
def forward(self, x):
x = self.cv1(x)
return self.cv2(torch.cat([x] + [m(x) for m in self.m], dim=1))
class Focus(nn.Module):
# Focus wh information into c-space
def __init__(self, in_channel, out_channel, kernel_size=1, stride=1, padding=None, groups=1, activation=True):
super(Focus, self).__init__()
self.conv = Conv(in_channel * 4, out_channel, kernel_size, stride, padding, groups, activation)
def forward(self, x): # x(b,c,w,h) -> y(b,4c,w/2,h/2)
return self.conv(torch.cat([x[..., ::2, ::2], x[..., 1::2, ::2], x[..., ::2, 1::2], x[..., 1::2, 1::2]], dim=1))
class Concat(nn.Module):
# Concatenate a list of tensors along dimension
def __init__(self, dimension=1):
super(Concat, self).__init__()
self.d = dimension
def forward(self, x):
return torch.cat(x, di
|
d)
class Detect(nn.Module):
def __init__(self, num_classes, num_anchor, reference_channels):
super(Detect, self).__init__()
self.num_anchor = num_anchor
self.num_classes = num_classes
self.num_output = self.num_classes + 5
self.m = nn.ModuleList(nn.Conv2d(input_channel, self.num_output * self.num_anchor, 1) for input_channel in reference_channels)
self.init_weight()
def forward(self, x):
for ilevel, module in enumerate(self.m):
x[ilevel] = module(x[ilevel])
return x
def init_weight(self):
strides = [8, 16, 32]
for head, stride in zip(self.m, strides):
bias = head.bias.view(self.num_anchor, -1)
bias[:, 4] += math.log(8 / (640 / stride) ** 2)
bias[:, 5:] += math.log(0.6 / (self.num_classes - 0.99))
head.bias = nn.Parameter(bias.view(-1), requires_grad=True)
class Yolo(nn.Module):
def __init__(self, num_classes, config_file, rank=0):
super(Yolo, self).__init__()
self.num_classes = num_classes
self.rank = rank
self.strides = [8, 16, 32]
self.model, self.saved_index, anchors = self.build_model(config_file)
self.register_buffer("anchors", torch.FloatTensor(anchors).view(3, 3, 2) / torch.FloatTensor(self.strides).view(3, 1, 1))
self.apply(self.init_weight)
def set_new_anchors(self, anchors):
# 对设置的anchors缩放到特征图大小
self.anchors[...] = anchors / torch.FloatTensor(self.strides).view(3, 1, 1)
def init_weight(self, m):
type_t = type(m)
if type_t is nn.Conv2d:
# pass init
pass
elif type_t is nn.BatchNorm2d:
m.eps = 1e-3
m.momentum = 0.03
elif type_t in [nn.LeakyReLU, nn.ReLU, nn.ReLU6]:
m.inplace = True
def forward(self, x):
y = []
for module in self.model:
if module.from_index != -1:
if isinstance(module.from_index, int):
x = y[module.from_index]
else:
xout = []
for i in module.from_index:
if i == -1:
xval = x
else:
xval = y[i]
xout.append(xval)
x = xout
x = module(x)
y.append(x if module.layer_index in self.saved_index else None)
return x
def parse_string(self, value):
if value == "None":
return None
elif value == "True":
return True
elif value == "False":
return False
else:
return value
def fuse(self): # fuse model Conv2d() + BatchNorm2d() layers
print('Fusing layers... ', end='')
for m in self.model.modules():
if type(m) is Conv:
m.conv = nn_utils.fuse_conv_and_bn(m.conv, m.bn) # update conv
m.bn = None # remove batchnorm
m.forward = m.fuse_forward # update forward
return self
def build_model(self, config_file, input_channel=3):
with open(config_file) as f:
self.yaml = yaml.load(f, Loader=yaml.FullLoader)
all_layers_cfg_list = self.yaml["backbone"] + self.yaml["head"]
anchors, depth_multiple, width_multiple = [self.yaml[item] for item in ["anchors", "depth_multiple", "width_multiple"]]
num_classes = self.num_classes
num_anchor = len(anchors[0]) // 2
num_output = num_anchor * (num_classes + 5)
all_layers_channels = [input_channel]
all_layers = []
saved_layer_index = []
for layer_index, (from_index, repeat_count, module_name, args) in enumerate(all_layers_cfg_list):
args = [self.parse_string(a) for a in args]
module_function = eval(module_name)
if repeat_count > 1:
repeat_count = max(round(repeat_count * depth_multiple), 1)
if module_function in [Conv, Bottleneck, SPP, Focus, BottleneckCSP]:
channel_input, channel_output = all_layers_channels[from_index], args[0]
if channel_output != num_output:
channel_output = make_divisible(channel_output * width_multiple, 8)
args = [channel_input, channel_output, *args[1:]]
if module_function in [BottleneckCSP]:
args.insert(2, repeat_count)
repeat_count = 1
elif module_function is Concat:
channel_output = sum([all_layers_channels[-1 if x == -1 else x + 1] for x in from_index])
|
m=self.
|
identifier_name
|
filetrace_test.go
|
0 exit_group(0) = ?`,
}
// TestStraceParse1 tests strace level1 parser (joining) by counting and
// checking strings.
func TestStraceParse1(t *testing.T) {
// Count strings that will be parsed away by StraceParser1
n := len(straceout)
for _, l := range straceout {
if strings.Contains(l, "resumed") || strings.Contains(l, "--- SIG") {
n--
continue
}
m, err := regexp.MatchString("^[0-9]+ \\?.*", l)
if err != nil {
t.Error(err)
}
if m {
n--
continue
}
}
if n == len(straceout) {
t.Error("test string has no level 1 parser tokens")
}
// Parse, and check that they went away and that the count is right
parsed := make([]string, 0, len(straceout))
for l := range StraceParse1(ChanFromList(straceout)) {
if strings.Contains(l, "resumed") || strings.Contains(l, "finished") {
t.Error("found invalid string in parsed results: " + l)
}
parsed = append(parsed, l)
}
if len(parsed) != n {
t.Error("incorrect len of parsed strings")
}
}
// TestStraceParse2Basic tests strace level2 parser by counting parsed entities.
func TestStraceParse2Basic(t *testing.T) {
nopen := 0
nexec := 0
for _, l := range straceout {
if strings.Contains(l, " open(") {
nopen++
}
if strings.Contains(l, " execve(") {
nexec++
}
}
syscalls := map[string]int{}
for info := range StraceParse2(StraceParse1(ChanFromList(straceout))) {
syscalls[info.syscall]++
}
if nopen != syscalls["open"] {
t.Errorf("\"open\" count mismatch: %d != %d", nopen, syscalls["open"])
}
if nexec != syscalls["execve"] {
t.Errorf("\"execve\" count mismatch: %d != %d", nexec, syscalls["execve"])
}
}
// TestStraceParse2Args tests strace level2 argument splitting.
func TestStraceParse2Args(t *testing.T) {
tests := []struct {
str string
ans []string
}{
{"asdf", []string{"asdf"}},
{"as, df", []string{"as", "df"}},
{"a {s, d} f", []string{"a {s, d} f"}},
{"{as, df}", []string{"{as, df}"}},
{`"as, df"`, []string{`"as, df"`}},
{`"as, df", gh`, []string{`"as, df"`, "gh"}},
{`"as, df\", gh"`, []string{`"as, df\", gh"`}},
{`"as, df\""`, []string{`"as, df\""`}},
}
for _, tst := range tests {
a := StraceParse2Argsplit(tst.str)
if !reflect.DeepEqual(a, tst.ans) {
t.Error(a, "!=", tst.ans)
}
}
}
// TestStraceParse2Lines tests a specific line-parsing.
func TestStraceParse2Lines(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
}()
for info := range StraceParse2(c) {
tests := []struct {
ok bool
str string
}{
{info.pid == 16821, "pid mismatch"},
{info.syscall == "open", "syscall mismatch"},
{info.result == 3, "result mismatch"},
{info.body == `"/etc/ld.so.cache", O_RDONLY|O_CLOEXEC`, "body mismatch"},
{info.err == "", "error mismatch"},
}
for _, tst := range tests {
if !tst.ok {
t.Error(tst.str)
}
}
ans := []string{
`"/etc/ld.so.cache"`,
`O_RDONLY|O_CLOEXEC`,
}
if len(ans) != len(info.args) {
t.Errorf("args len mismatch: len(%s)=%d != len(%s)=%d", info.args, len(info.args), ans, len(ans))
}
for i := 0; i < len(info.args); i++ {
if ans[i] != info.args[i] {
t.Errorf("arg %d mismatch", i)
}
}
}
}
// TestStraceParse3 tests StraceParse3 and StraceParse2
func TestStraceParse3(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
c <- `16821 open("w", O_WRONLY|O_CREAT|O_TRUNC|O_CLOEXEC) = 4`
c <- `16821 open("r", O_RDONLY|O_CLOEXEC) = 5`
c <- `16821 open("rw", O_RDWR|O_NONBLOCK) = 6`
c <- `16821 creat("c", 01) = 6`
}()
r, w := StraceParse3(StraceParse2(c), "")
rok := map[string]bool{
"/etc/ld.so.cache": true,
"r": true,
"rw": true,
}
if !reflect.DeepEqual(r, rok) {
t.Error(r, "!=", rok)
}
wok := map[string]bool{
"w": true,
"c": true,
"rw": true,
}
if !reflect.DeepEqual(w, wok) {
t.Error(w, "!=", wok)
}
}
// Test real applications:
// straceRbase has the base read files for the OS where the tests are run.
var straceRbase map[string]bool
// empty is an empty map
var empty = map[string]bool{}
// filetraceTest is the primitive test function that runs the provided command
// and checks if the set of files read and written match the ones provided.
func filetraceTest(t *testing.T, cmd string, dir string, rok map[string]bool, wok map[string]bool) {
if len(straceRbase) == 0 {
straceRbase, _ = FileTrace("", nil, "")
}
rt, wt := FileTrace(cmd, nil, dir)
rok2 := map[string]bool{}
for r := range rok {
rok2[r] = true
}
for r := range straceRbase {
rok2[r] = true
}
if !reflect.DeepEqual(rt, rok2) {
t.Error("r", rt, "!=", rok2)
}
if !reflect.DeepEqual(wt, wok) {
t.Error("w", wt, "!=", wok)
}
}
// TestFiletraceEchocat is the base test of read/write that runs an echo with the
// output redirected to a file, and a cat that reads that file.
func TestFiletraceEchocat(t *testing.T) {
empty := map[string]bool{}
filetraceTest(t,
"echo asdf > t",
"",
empty,
map[string]bool{"t": true})
defer func() {
if err := os.Remove("t"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cat t > h",
"",
map[string]bool{"t": true},
map[string]bool{"h": true})
defer func() {
if err := os.Remove("h"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cp t j",
"",
map[string]bool{"t": true},
map[string]bool{"j": true})
defer func() {
if err := os.Remove("j"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceChdir tests directory chaging.
func TestFiletraceChdir(t *testing.T) {
filetraceTest(t,
"mkdir d; cd d; echo asdf > t",
"",
empty,
map[string]bool{"d/t": true})
defer func() {
if err := os.Remove("d/t"); err != nil {
t.Error(err)
}
if err := os.Remove("d"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceEnv tests the environment argument.
func TestFiletraceEnv(t *testing.T) {
FileTrace("env > e.txt", map[string]string{"x": "y"}, "")
defer func() {
if err := os.Remove("e.txt"); err != nil
|
{
t.Error(err)
}
|
conditional_block
|
|
filetrace_test.go
|
explain"], [/* 34 vars */]) = -1 ENOENT (No such file or directory)`,
`16820 --- SIGCHLD (Child exited) @ 0 (0) ---`,
`16820 execve("/usr/bin/tclsh", ["tclsh", "/usr/bin/unbuffer", "scons", "--max-drift=1", "--implicit-cache", "--debug=explain"], [/* 34 vars */]) = 0`,
`16820 execve("/usr/bin/tclsh", ["tclsh", "/usr/bin/unbuffer", "scons", "--max-drift=1", "--implicit-cache", "--debug=explain"], [/* 34 vars */]) = ? <unavailable>`,
`16820 ????(= ? <unavailable>`,
`16821 open("/dev/tty", O_RDWR|O_NONBLOCK|O_LARGEFILE) = 3`,
`16820 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC <unfinished ...>`,
`16820 <... ???? resumed> ) = ? <unavailable>`,
`16821 open("/usr/lib/locale/locale-archive", O_RDONLY|O_LARGEFILE|O_CLOEXEC <unfinished ...>`,
`16820 <... open resumed> ) = 3`,
`16821 <... open resumed> ) = 3`,
`16820 open("/usr/lib\"/libtcl8.5.so.0", O_RDONLY|O_CLOEXEC) = 3`,
`16820 exit_group(0) = ?`,
}
// TestStraceParse1 tests strace level1 parser (joining) by counting and
// checking strings.
func TestStraceParse1(t *testing.T) {
// Count strings that will be parsed away by StraceParser1
n := len(straceout)
for _, l := range straceout {
if strings.Contains(l, "resumed") || strings.Contains(l, "--- SIG") {
n--
continue
}
m, err := regexp.MatchString("^[0-9]+ \\?.*", l)
if err != nil {
t.Error(err)
}
if m {
n--
continue
}
}
if n == len(straceout) {
t.Error("test string has no level 1 parser tokens")
}
// Parse, and check that they went away and that the count is right
parsed := make([]string, 0, len(straceout))
for l := range StraceParse1(ChanFromList(straceout)) {
if strings.Contains(l, "resumed") || strings.Contains(l, "finished") {
t.Error("found invalid string in parsed results: " + l)
}
parsed = append(parsed, l)
}
if len(parsed) != n {
t.Error("incorrect len of parsed strings")
}
}
// TestStraceParse2Basic tests strace level2 parser by counting parsed entities.
func TestStraceParse2Basic(t *testing.T) {
nopen := 0
nexec := 0
for _, l := range straceout {
if strings.Contains(l, " open(") {
nopen++
}
if strings.Contains(l, " execve(") {
nexec++
}
}
syscalls := map[string]int{}
for info := range StraceParse2(StraceParse1(ChanFromList(straceout))) {
syscalls[info.syscall]++
}
if nopen != syscalls["open"] {
t.Errorf("\"open\" count mismatch: %d != %d", nopen, syscalls["open"])
}
if nexec != syscalls["execve"] {
t.Errorf("\"execve\" count mismatch: %d != %d", nexec, syscalls["execve"])
}
}
// TestStraceParse2Args tests strace level2 argument splitting.
func TestStraceParse2Args(t *testing.T) {
tests := []struct {
str string
ans []string
}{
{"asdf", []string{"asdf"}},
{"as, df", []string{"as", "df"}},
{"a {s, d} f", []string{"a {s, d} f"}},
{"{as, df}", []string{"{as, df}"}},
{`"as, df"`, []string{`"as, df"`}},
{`"as, df", gh`, []string{`"as, df"`, "gh"}},
{`"as, df\", gh"`, []string{`"as, df\", gh"`}},
{`"as, df\""`, []string{`"as, df\""`}},
}
for _, tst := range tests {
a := StraceParse2Argsplit(tst.str)
if !reflect.DeepEqual(a, tst.ans) {
t.Error(a, "!=", tst.ans)
}
}
}
// TestStraceParse2Lines tests a specific line-parsing.
func TestStraceParse2Lines(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
}()
for info := range StraceParse2(c) {
tests := []struct {
ok bool
str string
}{
{info.pid == 16821, "pid mismatch"},
{info.syscall == "open", "syscall mismatch"},
{info.result == 3, "result mismatch"},
{info.body == `"/etc/ld.so.cache", O_RDONLY|O_CLOEXEC`, "body mismatch"},
{info.err == "", "error mismatch"},
}
for _, tst := range tests {
if !tst.ok {
t.Error(tst.str)
}
}
ans := []string{
`"/etc/ld.so.cache"`,
`O_RDONLY|O_CLOEXEC`,
}
if len(ans) != len(info.args) {
t.Errorf("args len mismatch: len(%s)=%d != len(%s)=%d", info.args, len(info.args), ans, len(ans))
}
for i := 0; i < len(info.args); i++ {
if ans[i] != info.args[i] {
t.Errorf("arg %d mismatch", i)
}
}
}
}
// TestStraceParse3 tests StraceParse3 and StraceParse2
func TestStraceParse3(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
c <- `16821 open("w", O_WRONLY|O_CREAT|O_TRUNC|O_CLOEXEC) = 4`
c <- `16821 open("r", O_RDONLY|O_CLOEXEC) = 5`
c <- `16821 open("rw", O_RDWR|O_NONBLOCK) = 6`
c <- `16821 creat("c", 01) = 6`
}()
r, w := StraceParse3(StraceParse2(c), "")
rok := map[string]bool{
"/etc/ld.so.cache": true,
"r": true,
"rw": true,
}
if !reflect.DeepEqual(r, rok) {
t.Error(r, "!=", rok)
}
wok := map[string]bool{
"w": true,
"c": true,
"rw": true,
}
if !reflect.DeepEqual(w, wok) {
t.Error(w, "!=", wok)
}
}
// Test real applications:
// straceRbase has the base read files for the OS where the tests are run.
var straceRbase map[string]bool
// empty is an empty map
var empty = map[string]bool{}
// filetraceTest is the primitive test function that runs the provided command
// and checks if the set of files read and written match the ones provided.
func filetraceTest(t *testing.T, cmd string, dir string, rok map[string]bool, wok map[string]bool) {
if len(straceRbase) == 0 {
straceRbase, _ = FileTrace("", nil, "")
}
rt, wt := FileTrace(cmd, nil, dir)
rok2 := map[string]bool{}
for r := range rok {
rok2[r] = true
}
for r := range straceRbase {
rok2[r] = true
}
if !reflect.DeepEqual(rt, rok2) {
t.Error("r", rt, "!=", rok2)
}
if !reflect.DeepEqual(wt, wok) {
t.Error("w", wt, "!=", wok)
}
}
// TestFiletraceEchocat is the base test of read/write that runs an echo with the
// output redirected to a file, and a cat that reads that file.
func
|
TestFiletraceEchocat
|
identifier_name
|
|
filetrace_test.go
|
}
// TestStraceParse2Basic tests strace level2 parser by counting parsed entities.
func TestStraceParse2Basic(t *testing.T) {
nopen := 0
nexec := 0
for _, l := range straceout {
if strings.Contains(l, " open(") {
nopen++
}
if strings.Contains(l, " execve(") {
nexec++
}
}
syscalls := map[string]int{}
for info := range StraceParse2(StraceParse1(ChanFromList(straceout))) {
syscalls[info.syscall]++
}
if nopen != syscalls["open"] {
t.Errorf("\"open\" count mismatch: %d != %d", nopen, syscalls["open"])
}
if nexec != syscalls["execve"] {
t.Errorf("\"execve\" count mismatch: %d != %d", nexec, syscalls["execve"])
}
}
// TestStraceParse2Args tests strace level2 argument splitting.
func TestStraceParse2Args(t *testing.T) {
tests := []struct {
str string
ans []string
}{
{"asdf", []string{"asdf"}},
{"as, df", []string{"as", "df"}},
{"a {s, d} f", []string{"a {s, d} f"}},
{"{as, df}", []string{"{as, df}"}},
{`"as, df"`, []string{`"as, df"`}},
{`"as, df", gh`, []string{`"as, df"`, "gh"}},
{`"as, df\", gh"`, []string{`"as, df\", gh"`}},
{`"as, df\""`, []string{`"as, df\""`}},
}
for _, tst := range tests {
a := StraceParse2Argsplit(tst.str)
if !reflect.DeepEqual(a, tst.ans) {
t.Error(a, "!=", tst.ans)
}
}
}
// TestStraceParse2Lines tests a specific line-parsing.
func TestStraceParse2Lines(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
}()
for info := range StraceParse2(c) {
tests := []struct {
ok bool
str string
}{
{info.pid == 16821, "pid mismatch"},
{info.syscall == "open", "syscall mismatch"},
{info.result == 3, "result mismatch"},
{info.body == `"/etc/ld.so.cache", O_RDONLY|O_CLOEXEC`, "body mismatch"},
{info.err == "", "error mismatch"},
}
for _, tst := range tests {
if !tst.ok {
t.Error(tst.str)
}
}
ans := []string{
`"/etc/ld.so.cache"`,
`O_RDONLY|O_CLOEXEC`,
}
if len(ans) != len(info.args) {
t.Errorf("args len mismatch: len(%s)=%d != len(%s)=%d", info.args, len(info.args), ans, len(ans))
}
for i := 0; i < len(info.args); i++ {
if ans[i] != info.args[i] {
t.Errorf("arg %d mismatch", i)
}
}
}
}
// TestStraceParse3 tests StraceParse3 and StraceParse2
func TestStraceParse3(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
c <- `16821 open("w", O_WRONLY|O_CREAT|O_TRUNC|O_CLOEXEC) = 4`
c <- `16821 open("r", O_RDONLY|O_CLOEXEC) = 5`
c <- `16821 open("rw", O_RDWR|O_NONBLOCK) = 6`
c <- `16821 creat("c", 01) = 6`
}()
r, w := StraceParse3(StraceParse2(c), "")
rok := map[string]bool{
"/etc/ld.so.cache": true,
"r": true,
"rw": true,
}
if !reflect.DeepEqual(r, rok) {
t.Error(r, "!=", rok)
}
wok := map[string]bool{
"w": true,
"c": true,
"rw": true,
}
if !reflect.DeepEqual(w, wok) {
t.Error(w, "!=", wok)
}
}
// Test real applications:
// straceRbase has the base read files for the OS where the tests are run.
var straceRbase map[string]bool
// empty is an empty map
var empty = map[string]bool{}
// filetraceTest is the primitive test function that runs the provided command
// and checks if the set of files read and written match the ones provided.
func filetraceTest(t *testing.T, cmd string, dir string, rok map[string]bool, wok map[string]bool) {
if len(straceRbase) == 0 {
straceRbase, _ = FileTrace("", nil, "")
}
rt, wt := FileTrace(cmd, nil, dir)
rok2 := map[string]bool{}
for r := range rok {
rok2[r] = true
}
for r := range straceRbase {
rok2[r] = true
}
if !reflect.DeepEqual(rt, rok2) {
t.Error("r", rt, "!=", rok2)
}
if !reflect.DeepEqual(wt, wok) {
t.Error("w", wt, "!=", wok)
}
}
// TestFiletraceEchocat is the base test of read/write that runs an echo with the
// output redirected to a file, and a cat that reads that file.
func TestFiletraceEchocat(t *testing.T) {
empty := map[string]bool{}
filetraceTest(t,
"echo asdf > t",
"",
empty,
map[string]bool{"t": true})
defer func() {
if err := os.Remove("t"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cat t > h",
"",
map[string]bool{"t": true},
map[string]bool{"h": true})
defer func() {
if err := os.Remove("h"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cp t j",
"",
map[string]bool{"t": true},
map[string]bool{"j": true})
defer func() {
if err := os.Remove("j"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceChdir tests directory chaging.
func TestFiletraceChdir(t *testing.T) {
filetraceTest(t,
"mkdir d; cd d; echo asdf > t",
"",
empty,
map[string]bool{"d/t": true})
defer func() {
if err := os.Remove("d/t"); err != nil {
t.Error(err)
}
if err := os.Remove("d"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceEnv tests the environment argument.
func TestFiletraceEnv(t *testing.T) {
FileTrace("env > e.txt", map[string]string{"x": "y"}, "")
defer func() {
if err := os.Remove("e.txt"); err != nil {
t.Error(err)
}
}()
data, err := ioutil.ReadFile("e.txt")
if err != nil {
t.Fatal(err)
}
datastr := string(data)
if !strings.Contains(datastr, "x=y") {
t.Fatalf("environment x=y not found in %s", datastr)
}
}
// TestFiletraceDir tests the dir argument.
func TestFiletraceDir(t *testing.T) {
os.Mkdir("d", 0755)
filetraceTest(t,
"mkdir -p s/ss; cd s; cd ss; echo asdf > t; echo zxcv > z; rm z",
"d",
empty,
map[string]bool{"s/ss/t": true})
defer func() {
for _, f := range []string{"d/s/ss/t", "d/s/ss", "d/s", "d"} {
if err := os.Remove(f); err != nil {
t.Error(err)
}
}
}()
}
// TestFiletraceRename tests renaming
func TestFiletraceRename(t *testing.T)
|
{
empty := map[string]bool{}
filetraceTest(t,
"echo asdf > t; mv t v",
"",
empty,
map[string]bool{"v": true})
defer func() {
if err := os.Remove("v"); err != nil {
t.Error(err)
}
}()
}
|
identifier_body
|
|
filetrace_test.go
|
([]string, 0, len(straceout))
for l := range StraceParse1(ChanFromList(straceout)) {
if strings.Contains(l, "resumed") || strings.Contains(l, "finished") {
t.Error("found invalid string in parsed results: " + l)
}
parsed = append(parsed, l)
}
if len(parsed) != n {
t.Error("incorrect len of parsed strings")
}
}
// TestStraceParse2Basic tests strace level2 parser by counting parsed entities.
func TestStraceParse2Basic(t *testing.T) {
nopen := 0
nexec := 0
for _, l := range straceout {
if strings.Contains(l, " open(") {
nopen++
}
if strings.Contains(l, " execve(") {
nexec++
}
}
syscalls := map[string]int{}
for info := range StraceParse2(StraceParse1(ChanFromList(straceout))) {
syscalls[info.syscall]++
}
if nopen != syscalls["open"] {
t.Errorf("\"open\" count mismatch: %d != %d", nopen, syscalls["open"])
}
if nexec != syscalls["execve"] {
t.Errorf("\"execve\" count mismatch: %d != %d", nexec, syscalls["execve"])
}
}
// TestStraceParse2Args tests strace level2 argument splitting.
func TestStraceParse2Args(t *testing.T) {
tests := []struct {
str string
ans []string
}{
{"asdf", []string{"asdf"}},
{"as, df", []string{"as", "df"}},
{"a {s, d} f", []string{"a {s, d} f"}},
{"{as, df}", []string{"{as, df}"}},
{`"as, df"`, []string{`"as, df"`}},
{`"as, df", gh`, []string{`"as, df"`, "gh"}},
{`"as, df\", gh"`, []string{`"as, df\", gh"`}},
{`"as, df\""`, []string{`"as, df\""`}},
}
for _, tst := range tests {
a := StraceParse2Argsplit(tst.str)
if !reflect.DeepEqual(a, tst.ans) {
t.Error(a, "!=", tst.ans)
}
}
}
// TestStraceParse2Lines tests a specific line-parsing.
func TestStraceParse2Lines(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
}()
for info := range StraceParse2(c) {
tests := []struct {
ok bool
str string
}{
{info.pid == 16821, "pid mismatch"},
{info.syscall == "open", "syscall mismatch"},
{info.result == 3, "result mismatch"},
{info.body == `"/etc/ld.so.cache", O_RDONLY|O_CLOEXEC`, "body mismatch"},
{info.err == "", "error mismatch"},
}
for _, tst := range tests {
if !tst.ok {
t.Error(tst.str)
}
}
ans := []string{
`"/etc/ld.so.cache"`,
`O_RDONLY|O_CLOEXEC`,
}
if len(ans) != len(info.args) {
t.Errorf("args len mismatch: len(%s)=%d != len(%s)=%d", info.args, len(info.args), ans, len(ans))
}
for i := 0; i < len(info.args); i++ {
if ans[i] != info.args[i] {
t.Errorf("arg %d mismatch", i)
}
}
}
}
// TestStraceParse3 tests StraceParse3 and StraceParse2
func TestStraceParse3(t *testing.T) {
c := make(chan string)
go func() {
defer close(c)
c <- `16821 open("/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3`
c <- `16821 open("w", O_WRONLY|O_CREAT|O_TRUNC|O_CLOEXEC) = 4`
c <- `16821 open("r", O_RDONLY|O_CLOEXEC) = 5`
c <- `16821 open("rw", O_RDWR|O_NONBLOCK) = 6`
c <- `16821 creat("c", 01) = 6`
}()
r, w := StraceParse3(StraceParse2(c), "")
rok := map[string]bool{
"/etc/ld.so.cache": true,
"r": true,
"rw": true,
}
if !reflect.DeepEqual(r, rok) {
t.Error(r, "!=", rok)
}
wok := map[string]bool{
"w": true,
"c": true,
"rw": true,
}
if !reflect.DeepEqual(w, wok) {
t.Error(w, "!=", wok)
}
}
// Test real applications:
// straceRbase has the base read files for the OS where the tests are run.
var straceRbase map[string]bool
// empty is an empty map
var empty = map[string]bool{}
// filetraceTest is the primitive test function that runs the provided command
// and checks if the set of files read and written match the ones provided.
func filetraceTest(t *testing.T, cmd string, dir string, rok map[string]bool, wok map[string]bool) {
if len(straceRbase) == 0 {
straceRbase, _ = FileTrace("", nil, "")
}
rt, wt := FileTrace(cmd, nil, dir)
rok2 := map[string]bool{}
for r := range rok {
rok2[r] = true
}
for r := range straceRbase {
rok2[r] = true
}
if !reflect.DeepEqual(rt, rok2) {
t.Error("r", rt, "!=", rok2)
}
if !reflect.DeepEqual(wt, wok) {
t.Error("w", wt, "!=", wok)
}
}
// TestFiletraceEchocat is the base test of read/write that runs an echo with the
// output redirected to a file, and a cat that reads that file.
func TestFiletraceEchocat(t *testing.T) {
empty := map[string]bool{}
filetraceTest(t,
"echo asdf > t",
"",
empty,
map[string]bool{"t": true})
defer func() {
if err := os.Remove("t"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cat t > h",
"",
map[string]bool{"t": true},
map[string]bool{"h": true})
defer func() {
if err := os.Remove("h"); err != nil {
t.Error(err)
}
}()
filetraceTest(t,
"cp t j",
"",
map[string]bool{"t": true},
map[string]bool{"j": true})
defer func() {
if err := os.Remove("j"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceChdir tests directory chaging.
func TestFiletraceChdir(t *testing.T) {
filetraceTest(t,
"mkdir d; cd d; echo asdf > t",
"",
empty,
map[string]bool{"d/t": true})
defer func() {
if err := os.Remove("d/t"); err != nil {
t.Error(err)
}
if err := os.Remove("d"); err != nil {
t.Error(err)
}
}()
}
// TestFiletraceEnv tests the environment argument.
func TestFiletraceEnv(t *testing.T) {
FileTrace("env > e.txt", map[string]string{"x": "y"}, "")
defer func() {
if err := os.Remove("e.txt"); err != nil {
t.Error(err)
}
}()
data, err := ioutil.ReadFile("e.txt")
if err != nil {
t.Fatal(err)
}
datastr := string(data)
if !strings.Contains(datastr, "x=y") {
t.Fatalf("environment x=y not found in %s", datastr)
}
}
// TestFiletraceDir tests the dir argument.
func TestFiletraceDir(t *testing.T) {
os.Mkdir("d", 0755)
filetraceTest(t,
"mkdir -p s/ss; cd s; cd ss; echo asdf > t; echo zxcv > z; rm z",
"d",
empty,
map[string]bool{"s/ss/t": true})
defer func() {
for _, f := range []string{"d/s/ss/t", "d/s/ss", "d/s", "d"} {
|
if err := os.Remove(f); err != nil {
t.Error(err)
}
|
random_line_split
|
|
settings.py
|
return git_hash.rstrip()
root = environ.Path(__file__) - 2 # two levels back in hierarchy
env = environ.Env(
DEBUG=(bool, False),
DJANGO_LOG_LEVEL=(str, "INFO"),
CONN_MAX_AGE=(int, 0),
SYSTEM_DATA_SOURCE_ID=(str, "hauki"),
LANGUAGES=(list, ["fi", "sv", "en"]),
DATABASE_URL=(str, "postgres:///hauki"),
TEST_DATABASE_URL=(str, ""),
TOKEN_AUTH_ACCEPTED_AUDIENCE=(str, ""),
TOKEN_AUTH_SHARED_SECRET=(str, ""),
SECRET_KEY=(str, ""),
ALLOWED_HOSTS=(list, []),
ADMINS=(list, []),
SECURE_PROXY_SSL_HEADER=(tuple, None),
MEDIA_ROOT=(environ.Path(), root("media")),
STATIC_ROOT=(environ.Path(), root("static")),
MEDIA_URL=(str, "/media/"),
STATIC_URL=(str, "/static/"),
TRUST_X_FORWARDED_HOST=(bool, False),
SENTRY_DSN=(str, ""),
SENTRY_ENVIRONMENT=(str, "development"),
COOKIE_PREFIX=(str, "hauki"),
INTERNAL_IPS=(list, []),
INSTANCE_NAME=(str, "Hauki"),
EXTRA_INSTALLED_APPS=(list, []),
ENABLE_DJANGO_EXTENSIONS=(bool, False),
MAIL_MAILGUN_KEY=(str, ""),
MAIL_MAILGUN_DOMAIN=(str, ""),
MAIL_MAILGUN_API=(str, ""),
RESOURCE_DEFAULT_TIMEZONE=(str, None),
)
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = root()
# Django environ has a nasty habit of complanining at level
# WARN about env file not being preset. Here we pre-empt it.
env_file_path = os.path.join(BASE_DIR, CONFIG_FILE_NAME)
if os.path.exists(env_file_path):
# Logging configuration is not available at this point
print(f"Reading config from {env_file_path}")
environ.Env.read_env(env_file_path)
DEBUG = env("DEBUG")
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = env("ALLOWED_HOSTS")
ADMINS = env("ADMINS")
INTERNAL_IPS = env("INTERNAL_IPS", default=(["127.0.0.1"] if DEBUG else []))
DATABASES = {"default": env.db()}
DATABASES["default"]["CONN_MAX_AGE"] = env("CONN_MAX_AGE")
if env("TEST_DATABASE_URL"):
DATABASES["default"]["TEST"] = env.db("TEST_DATABASE_URL")
DEFAULT_AUTO_FIELD = "django.db.models.AutoField"
AUTH_USER_MODEL = "users.User"
LOGIN_URL = "/login/"
LOGIN_REDIRECT_URL = "/v1/"
LOGOUT_REDIRECT_URL = "/v1/"
RESOURCE_DEFAULT_TIMEZONE = env("RESOURCE_DEFAULT_TIMEZONE")
DJANGO_ORGHIERARCHY_DATASOURCE_MODEL = "hours.DataSource"
SYSTEM_DATA_SOURCE_ID = env("SYSTEM_DATA_SOURCE_ID")
SITE_ID = 1
LOGGING = {
"version": 1,
"disable_existing_loggers": False,
"formatters": {
"timestamped_named": {
"format": "%(asctime)s %(name)s %(levelname)s: %(message)s",
},
},
"handlers": {
"console": {
"class": "logging.StreamHandler",
"formatter": "timestamped_named",
},
# Just for reference, not used
"blackhole": {
"class": "logging.NullHandler",
},
},
"loggers": {
"": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
"django": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
},
}
# Application definition
INSTALLED_APPS = [
"helusers.apps.HelusersConfig",
"modeltranslation",
"helusers.apps.HelusersAdminConfig",
"django.contrib.auth",
"django.contrib.contenttypes",
"django.contrib.sessions",
"django.contrib.messages",
"django.contrib.humanize",
"simple_history",
# disable Django’s development server static file handling
"whitenoise.runserver_nostatic",
"django.contrib.staticfiles",
"rest_framework",
"rest_framework.authtoken",
"django_filters",
"django_orghierarchy",
"timezone_field",
"mptt",
# Apps within this repository
"users",
"hours",
# OpenAPI
"drf_spectacular",
] + env("EXTRA_INSTALLED_APPS")
if env("SENTRY_DSN"):
sentry_sdk.init(
dsn=env("SENTRY_DSN"),
environment=env("SENTRY_ENVIRONMENT"),
release=get_git_revision_hash(),
integrations=[DjangoIntegration()],
)
MIDDLEWARE = [
# CorsMiddleware should be placed as high as possible and above WhiteNoiseMiddleware
# in particular
"corsheaders.middleware.CorsMiddleware",
# Ditto for securitymiddleware
"django.middleware.security.SecurityMiddleware",
"whitenoise.middleware.WhiteNoiseMiddleware",
"django.contrib.sessions.middleware.SessionMiddleware",
"django.middleware.common.CommonMiddleware",
"django.middleware.csrf.CsrfViewMiddleware",
"django.contrib.auth.middleware.AuthenticationMiddleware",
"django.contrib.messages.middleware.MessageMiddleware",
"django.middleware.clickjacking.XFrameOptionsMiddleware",
"simple_history.middleware.HistoryRequestMiddleware",
]
# django-extensions is a set of developer friendly tools
if env("ENABLE_DJANGO_EXTENSIONS"):
INSTALLED_APPS.append("django_extensions")
ROOT_URLCONF = "hauki.urls"
TEMPLATES = [
{
"BACKEND": "django.template.backends.django.DjangoTemplates",
"DIRS": [],
"APP_DIRS": True,
"OPTIONS": {
"context_processors": [
"django.template.context_processors.debug",
"django.template.context_processors.request",
"django.contrib.auth.context_processors.auth",
"django.contrib.messages.context_processors.messages",
],
},
},
]
WSGI_APPLICATION = "hauki.wsgi.application"
# Password validation
# https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
"NAME": "django.contrib.auth.password_validation."
"UserAttributeSimilarityValidator",
},
{
"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
},
{
"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
},
{
"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.0/topics/i18n/
# Map language codes to the (code, name) tuples used by Django
# We want to keep the ordering in LANGUAGES configuration variable,
# thus some gyrations
language_map = {x: y for x, y in GLOBAL_LANGUAGES}
try:
LANGUAGES = tuple((lang, language_map[lang]) for lang in env("LANGUAGES"))
except KeyError as e:
raise ImproperlyConfigured(f'unknown language code "{e.args[0]}"')
LANGUAGE_CODE = env("LANGUAGES")[0]
TIME_ZONE = "Europe/Helsinki"
USE_I18N = True
USE_L10N = True
USE_TZ = True
LOCALE_PATHS = [root("locale")]
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.0/howto/static-files/
STATIC_URL = env("STATIC_URL")
MEDIA_URL = env("MEDIA_URL")
STATIC_ROOT = env("STATIC_ROOT")
MEDIA_ROOT = env("MEDIA_ROOT")
# Whether to trust X-Forwarded-Host headers for all purposes
# where Django would need to make use of its own hostname
# fe. generating absolute URLs pointing to itself
# Most often used in reverse proxy setups
# https://docs.djangoproject.com/en/3.0/ref/settings/#use-x-forwarded-host
USE_X_FORWARDED_HOST = env("TRUST_X_FORWARDED_HOST")
# Specifies a header that is trusted to indicate that the request was using
# https while traversing over the Internet at large. This is used when
# a proxy terminates the TLS connection and forwards the request over
# a secure network. Specified using a tuple.
# https://docs.djangoproject.com/en/3.0/ref/settings/#secure-proxy-ssl-header
SECURE_PROXY_SSL_HEADER = env("SECURE_PROXY_SSL_HEADER")
CORS_ORIGIN_ALLOW_ALL = True
CSRF_COOKIE_NAME = "%s-csrftoken" % env("COOKIE_PREFIX")
SESSION_COOKIE_NAME = "%s-sessionid" % env("COOKIE_PREFIX")
# DRF Settings
# https://www.django-rest-framework.org/api-guide/settings/
REST
|
"""
Retrieve the git hash for the underlying git repository or die trying
We need a way to retrieve git revision hash for sentry reports
I assume that if we have a git repository available we will
have git-the-comamand as well
"""
try:
# We are not interested in gits complaints
git_hash = subprocess.check_output(
["git", "rev-parse", "HEAD"], stderr=subprocess.DEVNULL, encoding="utf8"
)
# ie. "git" was not found
# should we return a more generic meta hash here?
# like "undefined"?
except FileNotFoundError:
git_hash = "git_not_available"
except subprocess.CalledProcessError:
# Ditto
git_hash = "no_repository"
|
identifier_body
|
|
settings.py
|
except FileNotFoundError:
git_hash = "git_not_available"
except subprocess.CalledProcessError:
# Ditto
git_hash = "no_repository"
return git_hash.rstrip()
root = environ.Path(__file__) - 2 # two levels back in hierarchy
env = environ.Env(
DEBUG=(bool, False),
DJANGO_LOG_LEVEL=(str, "INFO"),
CONN_MAX_AGE=(int, 0),
SYSTEM_DATA_SOURCE_ID=(str, "hauki"),
LANGUAGES=(list, ["fi", "sv", "en"]),
DATABASE_URL=(str, "postgres:///hauki"),
TEST_DATABASE_URL=(str, ""),
TOKEN_AUTH_ACCEPTED_AUDIENCE=(str, ""),
TOKEN_AUTH_SHARED_SECRET=(str, ""),
SECRET_KEY=(str, ""),
ALLOWED_HOSTS=(list, []),
ADMINS=(list, []),
SECURE_PROXY_SSL_HEADER=(tuple, None),
MEDIA_ROOT=(environ.Path(), root("media")),
STATIC_ROOT=(environ.Path(), root("static")),
MEDIA_URL=(str, "/media/"),
STATIC_URL=(str, "/static/"),
TRUST_X_FORWARDED_HOST=(bool, False),
SENTRY_DSN=(str, ""),
SENTRY_ENVIRONMENT=(str, "development"),
COOKIE_PREFIX=(str, "hauki"),
INTERNAL_IPS=(list, []),
INSTANCE_NAME=(str, "Hauki"),
EXTRA_INSTALLED_APPS=(list, []),
ENABLE_DJANGO_EXTENSIONS=(bool, False),
MAIL_MAILGUN_KEY=(str, ""),
MAIL_MAILGUN_DOMAIN=(str, ""),
MAIL_MAILGUN_API=(str, ""),
RESOURCE_DEFAULT_TIMEZONE=(str, None),
)
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = root()
# Django environ has a nasty habit of complanining at level
# WARN about env file not being preset. Here we pre-empt it.
env_file_path = os.path.join(BASE_DIR, CONFIG_FILE_NAME)
if os.path.exists(env_file_path):
# Logging configuration is not available at this point
print(f"Reading config from {env_file_path}")
environ.Env.read_env(env_file_path)
DEBUG = env("DEBUG")
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = env("ALLOWED_HOSTS")
ADMINS = env("ADMINS")
INTERNAL_IPS = env("INTERNAL_IPS", default=(["127.0.0.1"] if DEBUG else []))
DATABASES = {"default": env.db()}
DATABASES["default"]["CONN_MAX_AGE"] = env("CONN_MAX_AGE")
if env("TEST_DATABASE_URL"):
DATABASES["default"]["TEST"] = env.db("TEST_DATABASE_URL")
DEFAULT_AUTO_FIELD = "django.db.models.AutoField"
AUTH_USER_MODEL = "users.User"
LOGIN_URL = "/login/"
LOGIN_REDIRECT_URL = "/v1/"
LOGOUT_REDIRECT_URL = "/v1/"
RESOURCE_DEFAULT_TIMEZONE = env("RESOURCE_DEFAULT_TIMEZONE")
DJANGO_ORGHIERARCHY_DATASOURCE_MODEL = "hours.DataSource"
SYSTEM_DATA_SOURCE_ID = env("SYSTEM_DATA_SOURCE_ID")
SITE_ID = 1
LOGGING = {
"version": 1,
"disable_existing_loggers": False,
"formatters": {
"timestamped_named": {
"format": "%(asctime)s %(name)s %(levelname)s: %(message)s",
},
},
"handlers": {
"console": {
"class": "logging.StreamHandler",
"formatter": "timestamped_named",
},
# Just for reference, not used
"blackhole": {
"class": "logging.NullHandler",
},
},
"loggers": {
"": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
"django": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
},
}
# Application definition
INSTALLED_APPS = [
"helusers.apps.HelusersConfig",
"modeltranslation",
"helusers.apps.HelusersAdminConfig",
"django.contrib.auth",
"django.contrib.contenttypes",
"django.contrib.sessions",
"django.contrib.messages",
"django.contrib.humanize",
"simple_history",
# disable Django’s development server static file handling
"whitenoise.runserver_nostatic",
"django.contrib.staticfiles",
"rest_framework",
"rest_framework.authtoken",
"django_filters",
"django_orghierarchy",
"timezone_field",
"mptt",
# Apps within this repository
"users",
"hours",
# OpenAPI
"drf_spectacular",
] + env("EXTRA_INSTALLED_APPS")
if env("SENTRY_DSN"):
se
|
MIDDLEWARE = [
# CorsMiddleware should be placed as high as possible and above WhiteNoiseMiddleware
# in particular
"corsheaders.middleware.CorsMiddleware",
# Ditto for securitymiddleware
"django.middleware.security.SecurityMiddleware",
"whitenoise.middleware.WhiteNoiseMiddleware",
"django.contrib.sessions.middleware.SessionMiddleware",
"django.middleware.common.CommonMiddleware",
"django.middleware.csrf.CsrfViewMiddleware",
"django.contrib.auth.middleware.AuthenticationMiddleware",
"django.contrib.messages.middleware.MessageMiddleware",
"django.middleware.clickjacking.XFrameOptionsMiddleware",
"simple_history.middleware.HistoryRequestMiddleware",
]
# django-extensions is a set of developer friendly tools
if env("ENABLE_DJANGO_EXTENSIONS"):
INSTALLED_APPS.append("django_extensions")
ROOT_URLCONF = "hauki.urls"
TEMPLATES = [
{
"BACKEND": "django.template.backends.django.DjangoTemplates",
"DIRS": [],
"APP_DIRS": True,
"OPTIONS": {
"context_processors": [
"django.template.context_processors.debug",
"django.template.context_processors.request",
"django.contrib.auth.context_processors.auth",
"django.contrib.messages.context_processors.messages",
],
},
},
]
WSGI_APPLICATION = "hauki.wsgi.application"
# Password validation
# https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
"NAME": "django.contrib.auth.password_validation."
"UserAttributeSimilarityValidator",
},
{
"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
},
{
"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
},
{
"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.0/topics/i18n/
# Map language codes to the (code, name) tuples used by Django
# We want to keep the ordering in LANGUAGES configuration variable,
# thus some gyrations
language_map = {x: y for x, y in GLOBAL_LANGUAGES}
try:
LANGUAGES = tuple((lang, language_map[lang]) for lang in env("LANGUAGES"))
except KeyError as e:
raise ImproperlyConfigured(f'unknown language code "{e.args[0]}"')
LANGUAGE_CODE = env("LANGUAGES")[0]
TIME_ZONE = "Europe/Helsinki"
USE_I18N = True
USE_L10N = True
USE_TZ = True
LOCALE_PATHS = [root("locale")]
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.0/howto/static-files/
STATIC_URL = env("STATIC_URL")
MEDIA_URL = env("MEDIA_URL")
STATIC_ROOT = env("STATIC_ROOT")
MEDIA_ROOT = env("MEDIA_ROOT")
# Whether to trust X-Forwarded-Host headers for all purposes
# where Django would need to make use of its own hostname
# fe. generating absolute URLs pointing to itself
# Most often used in reverse proxy setups
# https://docs.djangoproject.com/en/3.0/ref/settings/#use-x-forwarded-host
USE_X_FORWARDED_HOST = env("TRUST_X_FORWARDED_HOST")
# Specifies a header that is trusted to indicate that the request was using
# https while traversing over the Internet at large. This is used when
# a proxy terminates the TLS connection and forwards the request over
# a secure network. Specified using a tuple.
# https://docs.djangoproject.com/en/3.0/ref/settings/#secure-proxy-ssl-header
SECURE_PROXY_SSL_HEADER = env("SECURE_PROXY_SSL_HEADER")
CORS_ORIGIN_ALLOW_ALL = True
CSRF_COOKIE_NAME = "%s-csrftoken" % env("COOKIE_PREFIX")
SESSION_COOKIE_NAME = "%s-sessionid" % env("COOKIE_PREFIX")
# DRF Settings
# https://www.django-rest-framework.org/api-guide/settings/
REST_FRAMEWORK = {
"DEFAULT_RENDERER_CLASSES": [
"rest_framework.renderers.JSONRenderer",
"hours.renderers.BrowsableAPIRendererWithoutForms",
],
"DEFAULT_FILTER_BACKENDS": [
"rest_framework.filters.OrderingFilter",
"django_filters.rest_framework.DjangoFilterBackend",
],
"DEFAULT_AUTHENTICATION_CLASSES": [
"hours.authentication.HaukiSignedAuthentication",
"hours.authentication.HaukiTokenAuthentication",
"rest_framework.authentication.SessionAuthentication",
],
"DEFAULT_PERMISSION_CLASSES": [
"rest_framework.permissions.IsAuthenticatedOrReadOnly",
],
"DEFAULT_METADATA_CLASS": "hours.metadata.Trans
|
ntry_sdk.init(
dsn=env("SENTRY_DSN"),
environment=env("SENTRY_ENVIRONMENT"),
release=get_git_revision_hash(),
integrations=[DjangoIntegration()],
)
|
conditional_block
|
settings.py
|
except FileNotFoundError:
git_hash = "git_not_available"
except subprocess.CalledProcessError:
# Ditto
git_hash = "no_repository"
return git_hash.rstrip()
root = environ.Path(__file__) - 2 # two levels back in hierarchy
env = environ.Env(
DEBUG=(bool, False),
DJANGO_LOG_LEVEL=(str, "INFO"),
CONN_MAX_AGE=(int, 0),
SYSTEM_DATA_SOURCE_ID=(str, "hauki"),
LANGUAGES=(list, ["fi", "sv", "en"]),
DATABASE_URL=(str, "postgres:///hauki"),
TEST_DATABASE_URL=(str, ""),
TOKEN_AUTH_ACCEPTED_AUDIENCE=(str, ""),
TOKEN_AUTH_SHARED_SECRET=(str, ""),
SECRET_KEY=(str, ""),
ALLOWED_HOSTS=(list, []),
ADMINS=(list, []),
SECURE_PROXY_SSL_HEADER=(tuple, None),
MEDIA_ROOT=(environ.Path(), root("media")),
STATIC_ROOT=(environ.Path(), root("static")),
MEDIA_URL=(str, "/media/"),
STATIC_URL=(str, "/static/"),
TRUST_X_FORWARDED_HOST=(bool, False),
SENTRY_DSN=(str, ""),
SENTRY_ENVIRONMENT=(str, "development"),
COOKIE_PREFIX=(str, "hauki"),
INTERNAL_IPS=(list, []),
INSTANCE_NAME=(str, "Hauki"),
EXTRA_INSTALLED_APPS=(list, []),
ENABLE_DJANGO_EXTENSIONS=(bool, False),
MAIL_MAILGUN_KEY=(str, ""),
MAIL_MAILGUN_DOMAIN=(str, ""),
MAIL_MAILGUN_API=(str, ""),
RESOURCE_DEFAULT_TIMEZONE=(str, None),
)
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = root()
# Django environ has a nasty habit of complanining at level
# WARN about env file not being preset. Here we pre-empt it.
env_file_path = os.path.join(BASE_DIR, CONFIG_FILE_NAME)
if os.path.exists(env_file_path):
# Logging configuration is not available at this point
print(f"Reading config from {env_file_path}")
environ.Env.read_env(env_file_path)
DEBUG = env("DEBUG")
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = env("ALLOWED_HOSTS")
ADMINS = env("ADMINS")
INTERNAL_IPS = env("INTERNAL_IPS", default=(["127.0.0.1"] if DEBUG else []))
DATABASES = {"default": env.db()}
DATABASES["default"]["CONN_MAX_AGE"] = env("CONN_MAX_AGE")
if env("TEST_DATABASE_URL"):
DATABASES["default"]["TEST"] = env.db("TEST_DATABASE_URL")
DEFAULT_AUTO_FIELD = "django.db.models.AutoField"
AUTH_USER_MODEL = "users.User"
LOGIN_URL = "/login/"
LOGIN_REDIRECT_URL = "/v1/"
LOGOUT_REDIRECT_URL = "/v1/"
RESOURCE_DEFAULT_TIMEZONE = env("RESOURCE_DEFAULT_TIMEZONE")
DJANGO_ORGHIERARCHY_DATASOURCE_MODEL = "hours.DataSource"
SYSTEM_DATA_SOURCE_ID = env("SYSTEM_DATA_SOURCE_ID")
SITE_ID = 1
LOGGING = {
"version": 1,
"disable_existing_loggers": False,
"formatters": {
"timestamped_named": {
"format": "%(asctime)s %(name)s %(levelname)s: %(message)s",
},
},
"handlers": {
"console": {
"class": "logging.StreamHandler",
"formatter": "timestamped_named",
},
# Just for reference, not used
"blackhole": {
"class": "logging.NullHandler",
},
},
"loggers": {
"": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
"django": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
},
}
# Application definition
INSTALLED_APPS = [
"helusers.apps.HelusersConfig",
"modeltranslation",
"helusers.apps.HelusersAdminConfig",
"django.contrib.auth",
"django.contrib.contenttypes",
"django.contrib.sessions",
"django.contrib.messages",
"django.contrib.humanize",
"simple_history",
# disable Django’s development server static file handling
"whitenoise.runserver_nostatic",
"django.contrib.staticfiles",
"rest_framework",
"rest_framework.authtoken",
"django_filters",
"django_orghierarchy",
"timezone_field",
"mptt",
# Apps within this repository
"users",
"hours",
# OpenAPI
"drf_spectacular",
] + env("EXTRA_INSTALLED_APPS")
if env("SENTRY_DSN"):
sentry_sdk.init(
dsn=env("SENTRY_DSN"),
environment=env("SENTRY_ENVIRONMENT"),
release=get_git_revision_hash(),
integrations=[DjangoIntegration()],
)
MIDDLEWARE = [
# CorsMiddleware should be placed as high as possible and above WhiteNoiseMiddleware
# in particular
"corsheaders.middleware.CorsMiddleware",
# Ditto for securitymiddleware
"django.middleware.security.SecurityMiddleware",
"whitenoise.middleware.WhiteNoiseMiddleware",
"django.contrib.sessions.middleware.SessionMiddleware",
"django.middleware.common.CommonMiddleware",
"django.middleware.csrf.CsrfViewMiddleware",
"django.contrib.auth.middleware.AuthenticationMiddleware",
"django.contrib.messages.middleware.MessageMiddleware",
"django.middleware.clickjacking.XFrameOptionsMiddleware",
"simple_history.middleware.HistoryRequestMiddleware",
]
# django-extensions is a set of developer friendly tools
if env("ENABLE_DJANGO_EXTENSIONS"):
INSTALLED_APPS.append("django_extensions")
ROOT_URLCONF = "hauki.urls"
TEMPLATES = [
{
"BACKEND": "django.template.backends.django.DjangoTemplates",
"DIRS": [],
"APP_DIRS": True,
"OPTIONS": {
"context_processors": [
"django.template.context_processors.debug",
"django.template.context_processors.request",
"django.contrib.auth.context_processors.auth",
"django.contrib.messages.context_processors.messages",
],
},
},
]
WSGI_APPLICATION = "hauki.wsgi.application"
# Password validation
# https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
"NAME": "django.contrib.auth.password_validation."
"UserAttributeSimilarityValidator",
},
{
"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
},
{
"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
},
{
"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.0/topics/i18n/
# Map language codes to the (code, name) tuples used by Django
# We want to keep the ordering in LANGUAGES configuration variable,
# thus some gyrations
language_map = {x: y for x, y in GLOBAL_LANGUAGES}
try:
LANGUAGES = tuple((lang, language_map[lang]) for lang in env("LANGUAGES"))
except KeyError as e:
raise ImproperlyConfigured(f'unknown language code "{e.args[0]}"')
LANGUAGE_CODE = env("LANGUAGES")[0]
TIME_ZONE = "Europe/Helsinki"
USE_I18N = True
USE_L10N = True
USE_TZ = True
LOCALE_PATHS = [root("locale")]
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.0/howto/static-files/
|
STATIC_ROOT = env("STATIC_ROOT")
MEDIA_ROOT = env("MEDIA_ROOT")
# Whether to trust X-Forwarded-Host headers for all purposes
# where Django would need to make use of its own hostname
# fe. generating absolute URLs pointing to itself
# Most often used in reverse proxy setups
# https://docs.djangoproject.com/en/3.0/ref/settings/#use-x-forwarded-host
USE_X_FORWARDED_HOST = env("TRUST_X_FORWARDED_HOST")
# Specifies a header that is trusted to indicate that the request was using
# https while traversing over the Internet at large. This is used when
# a proxy terminates the TLS connection and forwards the request over
# a secure network. Specified using a tuple.
# https://docs.djangoproject.com/en/3.0/ref/settings/#secure-proxy-ssl-header
SECURE_PROXY_SSL_HEADER = env("SECURE_PROXY_SSL_HEADER")
CORS_ORIGIN_ALLOW_ALL = True
CSRF_COOKIE_NAME = "%s-csrftoken" % env("COOKIE_PREFIX")
SESSION_COOKIE_NAME = "%s-sessionid" % env("COOKIE_PREFIX")
# DRF Settings
# https://www.django-rest-framework.org/api-guide/settings/
REST_FRAMEWORK = {
"DEFAULT_RENDERER_CLASSES": [
"rest_framework.renderers.JSONRenderer",
"hours.renderers.BrowsableAPIRendererWithoutForms",
],
"DEFAULT_FILTER_BACKENDS": [
"rest_framework.filters.OrderingFilter",
"django_filters.rest_framework.DjangoFilterBackend",
],
"DEFAULT_AUTHENTICATION_CLASSES": [
"hours.authentication.HaukiSignedAuthentication",
"hours.authentication.HaukiTokenAuthentication",
"rest_framework.authentication.SessionAuthentication",
],
"DEFAULT_PERMISSION_CLASSES": [
"rest_framework.permissions.IsAuthenticatedOrReadOnly",
],
"DEFAULT_METADATA_CLASS": "hours.metadata.Trans
|
STATIC_URL = env("STATIC_URL")
MEDIA_URL = env("MEDIA_URL")
|
random_line_split
|
settings.py
|
() -> str:
"""
Retrieve the git hash for the underlying git repository or die trying
We need a way to retrieve git revision hash for sentry reports
I assume that if we have a git repository available we will
have git-the-comamand as well
"""
try:
# We are not interested in gits complaints
git_hash = subprocess.check_output(
["git", "rev-parse", "HEAD"], stderr=subprocess.DEVNULL, encoding="utf8"
)
# ie. "git" was not found
# should we return a more generic meta hash here?
# like "undefined"?
except FileNotFoundError:
git_hash = "git_not_available"
except subprocess.CalledProcessError:
# Ditto
git_hash = "no_repository"
return git_hash.rstrip()
root = environ.Path(__file__) - 2 # two levels back in hierarchy
env = environ.Env(
DEBUG=(bool, False),
DJANGO_LOG_LEVEL=(str, "INFO"),
CONN_MAX_AGE=(int, 0),
SYSTEM_DATA_SOURCE_ID=(str, "hauki"),
LANGUAGES=(list, ["fi", "sv", "en"]),
DATABASE_URL=(str, "postgres:///hauki"),
TEST_DATABASE_URL=(str, ""),
TOKEN_AUTH_ACCEPTED_AUDIENCE=(str, ""),
TOKEN_AUTH_SHARED_SECRET=(str, ""),
SECRET_KEY=(str, ""),
ALLOWED_HOSTS=(list, []),
ADMINS=(list, []),
SECURE_PROXY_SSL_HEADER=(tuple, None),
MEDIA_ROOT=(environ.Path(), root("media")),
STATIC_ROOT=(environ.Path(), root("static")),
MEDIA_URL=(str, "/media/"),
STATIC_URL=(str, "/static/"),
TRUST_X_FORWARDED_HOST=(bool, False),
SENTRY_DSN=(str, ""),
SENTRY_ENVIRONMENT=(str, "development"),
COOKIE_PREFIX=(str, "hauki"),
INTERNAL_IPS=(list, []),
INSTANCE_NAME=(str, "Hauki"),
EXTRA_INSTALLED_APPS=(list, []),
ENABLE_DJANGO_EXTENSIONS=(bool, False),
MAIL_MAILGUN_KEY=(str, ""),
MAIL_MAILGUN_DOMAIN=(str, ""),
MAIL_MAILGUN_API=(str, ""),
RESOURCE_DEFAULT_TIMEZONE=(str, None),
)
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = root()
# Django environ has a nasty habit of complanining at level
# WARN about env file not being preset. Here we pre-empt it.
env_file_path = os.path.join(BASE_DIR, CONFIG_FILE_NAME)
if os.path.exists(env_file_path):
# Logging configuration is not available at this point
print(f"Reading config from {env_file_path}")
environ.Env.read_env(env_file_path)
DEBUG = env("DEBUG")
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = env("ALLOWED_HOSTS")
ADMINS = env("ADMINS")
INTERNAL_IPS = env("INTERNAL_IPS", default=(["127.0.0.1"] if DEBUG else []))
DATABASES = {"default": env.db()}
DATABASES["default"]["CONN_MAX_AGE"] = env("CONN_MAX_AGE")
if env("TEST_DATABASE_URL"):
DATABASES["default"]["TEST"] = env.db("TEST_DATABASE_URL")
DEFAULT_AUTO_FIELD = "django.db.models.AutoField"
AUTH_USER_MODEL = "users.User"
LOGIN_URL = "/login/"
LOGIN_REDIRECT_URL = "/v1/"
LOGOUT_REDIRECT_URL = "/v1/"
RESOURCE_DEFAULT_TIMEZONE = env("RESOURCE_DEFAULT_TIMEZONE")
DJANGO_ORGHIERARCHY_DATASOURCE_MODEL = "hours.DataSource"
SYSTEM_DATA_SOURCE_ID = env("SYSTEM_DATA_SOURCE_ID")
SITE_ID = 1
LOGGING = {
"version": 1,
"disable_existing_loggers": False,
"formatters": {
"timestamped_named": {
"format": "%(asctime)s %(name)s %(levelname)s: %(message)s",
},
},
"handlers": {
"console": {
"class": "logging.StreamHandler",
"formatter": "timestamped_named",
},
# Just for reference, not used
"blackhole": {
"class": "logging.NullHandler",
},
},
"loggers": {
"": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
"django": {
"handlers": ["console"],
"level": os.getenv("DJANGO_LOG_LEVEL", "INFO"),
},
},
}
# Application definition
INSTALLED_APPS = [
"helusers.apps.HelusersConfig",
"modeltranslation",
"helusers.apps.HelusersAdminConfig",
"django.contrib.auth",
"django.contrib.contenttypes",
"django.contrib.sessions",
"django.contrib.messages",
"django.contrib.humanize",
"simple_history",
# disable Django’s development server static file handling
"whitenoise.runserver_nostatic",
"django.contrib.staticfiles",
"rest_framework",
"rest_framework.authtoken",
"django_filters",
"django_orghierarchy",
"timezone_field",
"mptt",
# Apps within this repository
"users",
"hours",
# OpenAPI
"drf_spectacular",
] + env("EXTRA_INSTALLED_APPS")
if env("SENTRY_DSN"):
sentry_sdk.init(
dsn=env("SENTRY_DSN"),
environment=env("SENTRY_ENVIRONMENT"),
release=get_git_revision_hash(),
integrations=[DjangoIntegration()],
)
MIDDLEWARE = [
# CorsMiddleware should be placed as high as possible and above WhiteNoiseMiddleware
# in particular
"corsheaders.middleware.CorsMiddleware",
# Ditto for securitymiddleware
"django.middleware.security.SecurityMiddleware",
"whitenoise.middleware.WhiteNoiseMiddleware",
"django.contrib.sessions.middleware.SessionMiddleware",
"django.middleware.common.CommonMiddleware",
"django.middleware.csrf.CsrfViewMiddleware",
"django.contrib.auth.middleware.AuthenticationMiddleware",
"django.contrib.messages.middleware.MessageMiddleware",
"django.middleware.clickjacking.XFrameOptionsMiddleware",
"simple_history.middleware.HistoryRequestMiddleware",
]
# django-extensions is a set of developer friendly tools
if env("ENABLE_DJANGO_EXTENSIONS"):
INSTALLED_APPS.append("django_extensions")
ROOT_URLCONF = "hauki.urls"
TEMPLATES = [
{
"BACKEND": "django.template.backends.django.DjangoTemplates",
"DIRS": [],
"APP_DIRS": True,
"OPTIONS": {
"context_processors": [
"django.template.context_processors.debug",
"django.template.context_processors.request",
"django.contrib.auth.context_processors.auth",
"django.contrib.messages.context_processors.messages",
],
},
},
]
WSGI_APPLICATION = "hauki.wsgi.application"
# Password validation
# https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
"NAME": "django.contrib.auth.password_validation."
"UserAttributeSimilarityValidator",
},
{
"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
},
{
"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
},
{
"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.0/topics/i18n/
# Map language codes to the (code, name) tuples used by Django
# We want to keep the ordering in LANGUAGES configuration variable,
# thus some gyrations
language_map = {x: y for x, y in GLOBAL_LANGUAGES}
try:
LANGUAGES = tuple((lang, language_map[lang]) for lang in env("LANGUAGES"))
except KeyError as e:
raise ImproperlyConfigured(f'unknown language code "{e.args[0]}"')
LANGUAGE_CODE = env("LANGUAGES")[0]
TIME_ZONE = "Europe/Helsinki"
USE_I18N = True
USE_L10N = True
USE_TZ = True
LOCALE_PATHS = [root("locale")]
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.0/howto/static-files/
STATIC_URL = env("STATIC_URL")
MEDIA_URL = env("MEDIA_URL")
STATIC_ROOT = env("STATIC_ROOT")
MEDIA_ROOT = env("MEDIA_ROOT")
# Whether to trust X-Forwarded-Host headers for all purposes
# where Django would need to make use of its own hostname
# fe. generating absolute URLs pointing to itself
# Most often used in reverse proxy setups
# https://docs.djangoproject.com/en/3.0/ref/settings/#use-x-forwarded-host
USE_X_FORWARDED_HOST = env("TRUST_X_FORWARDED_HOST")
# Specifies a header that is trusted to indicate that the request was using
# https while traversing over the Internet at large. This is used when
# a proxy terminates the TLS connection and forwards the request over
# a secure network. Specified using a tuple.
# https://docs.djangoproject.com/en/3.0/ref/settings/#secure-proxy-ssl-header
SECURE_PROXY_SSL_HEADER = env("SECURE_PROXY_SSL_HEADER")
CORS_ORIGIN_ALLOW_ALL = True
CSRF_COOKIE_NAME = "%s-csrftoken" % env("COOKIE_PREFIX")
SESSION_COOKIE_NAME = "%s-sessionid" % env("COOKIE_PREFIX")
# DRF Settings
# https://www.d
|
get_git_revision_hash
|
identifier_name
|
|
mod.rs
|
moves` closure which will be invoked with `(el, source,
/// handle, sibling)` whenever an element is clicked. If this closure returns
/// `false`, a drag event won't begin, and the event won't be prevented
/// either. The `handle` element will be the original click target, which
/// comes in handy to test if that element is an expected _"drag handle"_.
pub moves: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can set `accepts` to a closure with the following signature: `(el,
/// target, source, sibling)`. It'll be called to make sure that an element
/// `el`, that came from container `source`, can be dropped on container
/// `target` before a `sibling` element. The `sibling` can be `null`, which
/// would mean that the element would be placed as the last element in the
/// container. Note that if [`copy`](Options::copy) is set to `true`, `el` will be
/// set to the copy, instead of the originally dragged element.
pub accepts: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can provide an `invalid` closure with a `(el, handle)` signature.
/// This closure should return `true` for elements that shouldn't trigger a
/// drag. The `handle` argument is the element that was clicked, while `el`
/// is the item that would be dragged.
pub invalid: Box<dyn FnMut(JsValue, JsValue) -> bool>,
/// If `copy` is set to `true` _(or a closure that returns `true`)_, items
/// will be copied rather than moved. This implies the following differences:
///
/// Event | Move | Copy
/// ----------|------------------------------------------|---------------------------------------------
/// `drag` | Element will be concealed from `source` | Nothing happens
/// `drop` | Element will be moved into `target` | Element will be cloned into `target`
/// `remove` | Element will be removed from DOM | Nothing happens
/// `cancel` | Element will stay in `source` | Nothing happens
///
/// If a closure is passed, it'll be called whenever an element starts being
/// dragged in order to decide whether it should follow `copy` behavior or
/// not. This closure will be passed the element to be dragged as well as
/// its source container, in other words, the signature is `(el, handle)`.
///
/// `false` by default.
pub copy: CopyValue,
/// If [`copy`](Options::copy) is set to `true` _(or a closure that
/// returns `true`)_ and `copy_sort_source` is `true` as well, users will
/// be able to sort elements in `copy`-source containers.
///
/// `false` by default.
pub copy_sort_source: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `revert_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are moved back to the source element where
/// the drag event began, rather than stay at the _drop position previewed
/// by the feedback shadow_.
///
/// `false` by default.
pub revert_on_spill: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `remove_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are removed from the DOM. Note that `remove`
/// events won't fire if [`copy`](Options::copy) is set to `true`.
///
/// `false` by default.
pub remove_on_spill: bool,
/// When an element is dropped onto a container, it'll be placed near the
/// point where the mouse was released. If the `direction` is
/// [`Vertical`](Direction::Vertical),
/// the default value, the Y axis will be considered. Otherwise, if the
/// `direction` is [`Horizontal`](Direction::Horizontal),
/// the X axis will be considered.
///
/// [`Vertical`](Direction::Vertical), by default.
pub direction: Direction,
/// The DOM element where the mirror element displayed while dragging will
/// be appended to.
///
/// `document.body` by default.
pub mirror_container: JsValue,
/// When this option is enabled, if the user clicks on an input element the
/// drag won't start until their mouse pointer exits the input. This
/// translates into the user being able to select text in inputs contained
/// inside draggable elements, and still drag the element by moving their
/// mouse outside of the input -- so you get the best of both worlds.
///
/// `true` by default.
pub ignore_input_text_selection: bool,
/// The amount of horizontal movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_x: i32,
/// The amount of vertical movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_y: i32,
}
impl Default for Options {
fn default() -> Self {
Self {
is_container: Box::new(|_| false),
moves: Box::new(|_, _, _, _| true),
accepts: Box::new(|_, _, _, _| true),
invalid: Box::new(|_, _| false),
copy: CopyValue::Bool(false),
copy_sort_source: false,
revert_on_spill: false,
remove_on_spill: false,
direction: Direction::Vertical,
// Will default to document.body (avoiding web_sys dependency)
mirror_container: JsValue::UNDEFINED,
ignore_input_text_selection: true,
slide_factor_x: 0,
slide_factor_y: 0,
}
}
}
#[doc(hidden)]
#[wasm_bindgen]
pub struct OptionsImpl {
is_container_func: JsValue,
moves_func: JsValue,
accepts_func: JsValue,
invalid_func: JsValue,
copy_func_or_bool: JsValue,
#[wasm_bindgen(js_name = copySortSource)]
pub copy_sort_source: bool,
#[wasm_bindgen(js_name = revertOnSpill)]
pub revert_on_spill: bool,
#[wasm_bindgen(js_name = removeOnSpill)]
pub remove_on_spill: bool,
direction: String,
mirror_container_elem: JsValue,
#[wasm_bindgen(js_name = ignoreInputTextSelection)]
pub ignore_input_text_selection: bool,
#[wasm_bindgen(js_name = slideFactorX)]
pub slide_factor_x: i32,
#[wasm_bindgen(js_name = slideFactorY)]
pub slide_factor_y: i32,
}
impl From<Options> for OptionsImpl {
fn from(options: Options) -> Self {
OptionsImpl {
is_container_func: closure::to_js_1_ret(options.is_container),
moves_func: closure::to_js_4_ret(options.moves),
accepts_func: closure::to_js_4_ret(options.accepts),
invalid_func: closure::to_js_2_ret(options.invalid),
copy_func_or_bool: JsValue::from(options.copy),
mirror_container_elem: options.mirror_container,
copy_sort_source: options.copy_sort_source,
revert_on_spill: options.revert_on_spill,
remove_on_spill: options.remove_on_spill,
direction: options.direction.to_string(),
ignore_input_text_selection: options.ignore_input_text_selection,
slide_factor_x: options.slide_factor_x,
slide_factor_y: options.slide_factor_y,
}
}
}
impl Default for OptionsImpl {
fn default() -> Self {
OptionsImpl::from(Options::default())
}
}
#[wasm_bindgen]
#[doc(hidden)]
impl OptionsImpl {
#[wasm_bindgen(getter = isContainer)]
pub fn is_container_func(&self) -> JsValue {
self.is_container_func.clone()
}
#[wasm_bindgen(setter = isContainer)]
pub fn set_is_container_func(&mut self, val: JsValue) {
self.is_container_func = val;
}
#[wasm_bindgen(getter = moves)]
pub fn moves_func(&self) -> JsValue {
self.moves_func.clone()
}
#[wasm_bindgen(setter = moves)]
pub fn set_moves_func(&mut self, val: JsValue) {
self.moves_func = val;
}
#[wasm_bindgen(getter = accepts)]
pub fn accepts_func(&self) -> JsValue {
self.accepts_func.clone()
}
#[wasm_bindgen(setter = accepts)]
pub fn set_accepts_func(&mut self, val: JsValue) {
self.accepts_func = val;
}
#[wasm_bindgen(getter = invalid)]
pub fn invalid_func(&self) -> JsValue
|
{
self.invalid_func.clone()
}
|
identifier_body
|
|
mod.rs
|
: &str = "vertical";
const HORIZONTAL: &str = "horizontal";
match self {
Direction::Vertical => String::from(VERTICAL),
Direction::Horizontal => String::from(HORIZONTAL),
}
}
}
/// Used to pass options when activating Dragula
///
/// When passed to the [`dragula_options`](crate::dragula_options) function,
/// this struct can be used to specify options to control the behaviour of the
/// drag-and-drop functionality.
///
/// For example:
/// ```no_run
/// use dragula::*;
/// use dragula::options::CopyValue;
/// use web_sys::Element;
/// # use wasm_bindgen::JsValue;
///
/// # let element = JsValue::TRUE;
/// //--snip--
///
/// let options = Options {
/// invalid: Box::new(|el, _handle| {
/// Element::from(el).tag_name() == String::from("A")
/// }),
/// copy: CopyValue::Bool(true),
/// copy_sort_source: true,
/// remove_on_spill: true,
/// slide_factor_x: 10,
/// slide_factor_y: 10,
/// ..Options::default()
/// };
///
/// let drake = dragula_options(&[element], options);
///
/// //--snip--
/// ```
pub struct Options {
|
/// for this particular [`Drake`](crate::Drake) instance.
///
/// This closure will be invoked with the element that is being checked for
/// whether it is a container.
pub is_container: Box<dyn FnMut(JsValue) -> bool>,
/// You can define a `moves` closure which will be invoked with `(el, source,
/// handle, sibling)` whenever an element is clicked. If this closure returns
/// `false`, a drag event won't begin, and the event won't be prevented
/// either. The `handle` element will be the original click target, which
/// comes in handy to test if that element is an expected _"drag handle"_.
pub moves: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can set `accepts` to a closure with the following signature: `(el,
/// target, source, sibling)`. It'll be called to make sure that an element
/// `el`, that came from container `source`, can be dropped on container
/// `target` before a `sibling` element. The `sibling` can be `null`, which
/// would mean that the element would be placed as the last element in the
/// container. Note that if [`copy`](Options::copy) is set to `true`, `el` will be
/// set to the copy, instead of the originally dragged element.
pub accepts: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can provide an `invalid` closure with a `(el, handle)` signature.
/// This closure should return `true` for elements that shouldn't trigger a
/// drag. The `handle` argument is the element that was clicked, while `el`
/// is the item that would be dragged.
pub invalid: Box<dyn FnMut(JsValue, JsValue) -> bool>,
/// If `copy` is set to `true` _(or a closure that returns `true`)_, items
/// will be copied rather than moved. This implies the following differences:
///
/// Event | Move | Copy
/// ----------|------------------------------------------|---------------------------------------------
/// `drag` | Element will be concealed from `source` | Nothing happens
/// `drop` | Element will be moved into `target` | Element will be cloned into `target`
/// `remove` | Element will be removed from DOM | Nothing happens
/// `cancel` | Element will stay in `source` | Nothing happens
///
/// If a closure is passed, it'll be called whenever an element starts being
/// dragged in order to decide whether it should follow `copy` behavior or
/// not. This closure will be passed the element to be dragged as well as
/// its source container, in other words, the signature is `(el, handle)`.
///
/// `false` by default.
pub copy: CopyValue,
/// If [`copy`](Options::copy) is set to `true` _(or a closure that
/// returns `true`)_ and `copy_sort_source` is `true` as well, users will
/// be able to sort elements in `copy`-source containers.
///
/// `false` by default.
pub copy_sort_source: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `revert_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are moved back to the source element where
/// the drag event began, rather than stay at the _drop position previewed
/// by the feedback shadow_.
///
/// `false` by default.
pub revert_on_spill: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `remove_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are removed from the DOM. Note that `remove`
/// events won't fire if [`copy`](Options::copy) is set to `true`.
///
/// `false` by default.
pub remove_on_spill: bool,
/// When an element is dropped onto a container, it'll be placed near the
/// point where the mouse was released. If the `direction` is
/// [`Vertical`](Direction::Vertical),
/// the default value, the Y axis will be considered. Otherwise, if the
/// `direction` is [`Horizontal`](Direction::Horizontal),
/// the X axis will be considered.
///
/// [`Vertical`](Direction::Vertical), by default.
pub direction: Direction,
/// The DOM element where the mirror element displayed while dragging will
/// be appended to.
///
/// `document.body` by default.
pub mirror_container: JsValue,
/// When this option is enabled, if the user clicks on an input element the
/// drag won't start until their mouse pointer exits the input. This
/// translates into the user being able to select text in inputs contained
/// inside draggable elements, and still drag the element by moving their
/// mouse outside of the input -- so you get the best of both worlds.
///
/// `true` by default.
pub ignore_input_text_selection: bool,
/// The amount of horizontal movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_x: i32,
/// The amount of vertical movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_y: i32,
}
impl Default for Options {
fn default() -> Self {
Self {
is_container: Box::new(|_| false),
moves: Box::new(|_, _, _, _| true),
accepts: Box::new(|_, _, _, _| true),
invalid: Box::new(|_, _| false),
copy: CopyValue::Bool(false),
copy_sort_source: false,
revert_on_spill: false,
remove_on_spill: false,
direction: Direction::Vertical,
// Will default to document.body (avoiding web_sys dependency)
mirror_container: JsValue::UNDEFINED,
ignore_input_text_selection: true,
slide_factor_x: 0,
slide_factor_y: 0,
}
}
}
#[doc(hidden)]
#[wasm_bindgen]
pub struct OptionsImpl {
is_container_func: JsValue,
moves_func: JsValue,
accepts_func: JsValue,
invalid_func: JsValue,
copy_func_or_bool: JsValue,
#[wasm_bindgen(js_name = copySortSource)]
pub copy_sort_source: bool,
#[wasm_bindgen(js_name = revertOnSpill)]
pub revert_on_spill: bool,
#[wasm_bindgen(js_name = removeOnSpill)]
pub remove_on_spill: bool,
direction: String,
mirror_container_elem: JsValue,
#[wasm_bindgen(js_name = ignoreInputTextSelection)]
pub ignore_input_text_selection: bool,
#[wasm_bindgen(js_name = slideFactorX)]
pub slide_factor_x: i32,
#[wasm_bindgen(js_name = slideFactorY)]
pub slide_factor_y: i32,
}
impl From<Options> for OptionsImpl {
fn from(options: Options) -> Self {
OptionsImpl {
is_container_func: closure::to_js_1_ret(options.is_container),
moves_func: closure::to_js_4_ret(options.moves),
accepts_func: closure::to_js_4_ret(options.accepts),
invalid_func:
|
/// Besides the containers that you pass to [`dragula`](crate::dragula()),
/// or the containers you dynamically add, you can also use this closure to
/// specify any sort of logic that defines what is a container
|
random_line_split
|
mod.rs
|
;
/// use web_sys::Element;
/// # use wasm_bindgen::JsValue;
///
/// # let element = JsValue::TRUE;
/// //--snip--
///
/// let options = Options {
/// invalid: Box::new(|el, _handle| {
/// Element::from(el).tag_name() == String::from("A")
/// }),
/// copy: CopyValue::Bool(true),
/// copy_sort_source: true,
/// remove_on_spill: true,
/// slide_factor_x: 10,
/// slide_factor_y: 10,
/// ..Options::default()
/// };
///
/// let drake = dragula_options(&[element], options);
///
/// //--snip--
/// ```
pub struct Options {
/// Besides the containers that you pass to [`dragula`](crate::dragula()),
/// or the containers you dynamically add, you can also use this closure to
/// specify any sort of logic that defines what is a container
/// for this particular [`Drake`](crate::Drake) instance.
///
/// This closure will be invoked with the element that is being checked for
/// whether it is a container.
pub is_container: Box<dyn FnMut(JsValue) -> bool>,
/// You can define a `moves` closure which will be invoked with `(el, source,
/// handle, sibling)` whenever an element is clicked. If this closure returns
/// `false`, a drag event won't begin, and the event won't be prevented
/// either. The `handle` element will be the original click target, which
/// comes in handy to test if that element is an expected _"drag handle"_.
pub moves: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can set `accepts` to a closure with the following signature: `(el,
/// target, source, sibling)`. It'll be called to make sure that an element
/// `el`, that came from container `source`, can be dropped on container
/// `target` before a `sibling` element. The `sibling` can be `null`, which
/// would mean that the element would be placed as the last element in the
/// container. Note that if [`copy`](Options::copy) is set to `true`, `el` will be
/// set to the copy, instead of the originally dragged element.
pub accepts: Box<dyn FnMut(JsValue, JsValue, JsValue, JsValue) -> bool>,
/// You can provide an `invalid` closure with a `(el, handle)` signature.
/// This closure should return `true` for elements that shouldn't trigger a
/// drag. The `handle` argument is the element that was clicked, while `el`
/// is the item that would be dragged.
pub invalid: Box<dyn FnMut(JsValue, JsValue) -> bool>,
/// If `copy` is set to `true` _(or a closure that returns `true`)_, items
/// will be copied rather than moved. This implies the following differences:
///
/// Event | Move | Copy
/// ----------|------------------------------------------|---------------------------------------------
/// `drag` | Element will be concealed from `source` | Nothing happens
/// `drop` | Element will be moved into `target` | Element will be cloned into `target`
/// `remove` | Element will be removed from DOM | Nothing happens
/// `cancel` | Element will stay in `source` | Nothing happens
///
/// If a closure is passed, it'll be called whenever an element starts being
/// dragged in order to decide whether it should follow `copy` behavior or
/// not. This closure will be passed the element to be dragged as well as
/// its source container, in other words, the signature is `(el, handle)`.
///
/// `false` by default.
pub copy: CopyValue,
/// If [`copy`](Options::copy) is set to `true` _(or a closure that
/// returns `true`)_ and `copy_sort_source` is `true` as well, users will
/// be able to sort elements in `copy`-source containers.
///
/// `false` by default.
pub copy_sort_source: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `revert_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are moved back to the source element where
/// the drag event began, rather than stay at the _drop position previewed
/// by the feedback shadow_.
///
/// `false` by default.
pub revert_on_spill: bool,
/// By default, spilling an element outside of any containers will move the
/// element back to the _drop position previewed by the feedback shadow_.
/// Setting `remove_on_spill` to `true` will ensure elements dropped outside
/// of any approved containers are removed from the DOM. Note that `remove`
/// events won't fire if [`copy`](Options::copy) is set to `true`.
///
/// `false` by default.
pub remove_on_spill: bool,
/// When an element is dropped onto a container, it'll be placed near the
/// point where the mouse was released. If the `direction` is
/// [`Vertical`](Direction::Vertical),
/// the default value, the Y axis will be considered. Otherwise, if the
/// `direction` is [`Horizontal`](Direction::Horizontal),
/// the X axis will be considered.
///
/// [`Vertical`](Direction::Vertical), by default.
pub direction: Direction,
/// The DOM element where the mirror element displayed while dragging will
/// be appended to.
///
/// `document.body` by default.
pub mirror_container: JsValue,
/// When this option is enabled, if the user clicks on an input element the
/// drag won't start until their mouse pointer exits the input. This
/// translates into the user being able to select text in inputs contained
/// inside draggable elements, and still drag the element by moving their
/// mouse outside of the input -- so you get the best of both worlds.
///
/// `true` by default.
pub ignore_input_text_selection: bool,
/// The amount of horizontal movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_x: i32,
/// The amount of vertical movement (in pixels) for a click to be
/// considered a drag
///
/// `0` by default.
pub slide_factor_y: i32,
}
impl Default for Options {
fn default() -> Self {
Self {
is_container: Box::new(|_| false),
moves: Box::new(|_, _, _, _| true),
accepts: Box::new(|_, _, _, _| true),
invalid: Box::new(|_, _| false),
copy: CopyValue::Bool(false),
copy_sort_source: false,
revert_on_spill: false,
remove_on_spill: false,
direction: Direction::Vertical,
// Will default to document.body (avoiding web_sys dependency)
mirror_container: JsValue::UNDEFINED,
ignore_input_text_selection: true,
slide_factor_x: 0,
slide_factor_y: 0,
}
}
}
#[doc(hidden)]
#[wasm_bindgen]
pub struct OptionsImpl {
is_container_func: JsValue,
moves_func: JsValue,
accepts_func: JsValue,
invalid_func: JsValue,
copy_func_or_bool: JsValue,
#[wasm_bindgen(js_name = copySortSource)]
pub copy_sort_source: bool,
#[wasm_bindgen(js_name = revertOnSpill)]
pub revert_on_spill: bool,
#[wasm_bindgen(js_name = removeOnSpill)]
pub remove_on_spill: bool,
direction: String,
mirror_container_elem: JsValue,
#[wasm_bindgen(js_name = ignoreInputTextSelection)]
pub ignore_input_text_selection: bool,
#[wasm_bindgen(js_name = slideFactorX)]
pub slide_factor_x: i32,
#[wasm_bindgen(js_name = slideFactorY)]
pub slide_factor_y: i32,
}
impl From<Options> for OptionsImpl {
fn from(options: Options) -> Self {
OptionsImpl {
is_container_func: closure::to_js_1_ret(options.is_container),
moves_func: closure::to_js_4_ret(options.moves),
accepts_func: closure::to_js_4_ret(options.accepts),
invalid_func: closure::to_js_2_ret(options.invalid),
copy_func_or_bool: JsValue::from(options.copy),
mirror_container_elem: options.mirror_container,
copy_sort_source: options.copy_sort_source,
revert_on_spill: options.revert_on_spill,
remove_on_spill: options.remove_on_spill,
direction: options.direction.to_string(),
ignore_input_text_selection: options.ignore_input_text_selection,
slide_factor_x: options.slide_factor_x,
slide_factor_y: options.slide_factor_y,
}
}
}
impl Default for OptionsImpl {
fn
|
default
|
identifier_name
|
|
train.py
|
": [neighbor_features],
"features": [NODE_FEATURES],
"combined": [neighbor_features, NODE_FEATURES],
}
return dict(y for x in all_features[feat_type] for y in x.items())
class
|
:
def __init__(self, products_path, dataset_path, conf, logger, data_logger=None):
self.conf = conf
self._logger = logger
self._data_logger = EmptyLogger() if data_logger is None else data_logger
self.products_path = products_path
self.loader = GraphLoader(dataset_path, is_max_connected=False, norm_adj=conf["norm_adj"],
cuda_num=conf["cuda"], logger=self._logger)
self._criterion = torch.nn.NLLLoss()
def _get_models(self):
bow_feat = self.loader.bow_mx
topo_feat = self.loader.topo_mx
model1 = GCN(nfeat=bow_feat.shape[1],
hlayers=[self.conf["kipf"]["hidden"]],
nclass=self.loader.num_labels,
dropout=self.conf["kipf"]["dropout"])
opt1 = optim.Adam(model1.parameters(), lr=self.conf["kipf"]["lr"],
weight_decay=self.conf["kipf"]["weight_decay"])
model2 = GCNCombined(nbow=bow_feat.shape[1],
nfeat=topo_feat.shape[1],
hlayers=self.conf["hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"])
opt2 = optim.Adam(model2.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model3 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=None)
opt3 = optim.Adam(model3.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model4 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=AsymmetricGCN)
opt4 = optim.Adam(model4.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
return {
"kipf": {
"model": model1, "optimizer": opt1,
"arguments": [self.loader.bow_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"our_combined": {
"model": model2, "optimizer": opt2,
"arguments": [self.loader.bow_mx, self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
"topo_sym": {
"model": model3, "optimizer": opt3,
"arguments": [self.loader.topo_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"topo_asym": {
"model": model4, "optimizer": opt4,
"arguments": [self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
}
def run(self, train_p, feat_type):
features_meta = get_features(feat_type, is_directed=self.loader.is_graph_directed)
self.loader.split_train(train_p, features_meta)
models = self._get_models()
if self.conf["cuda"] is not None:
[model["model"].cuda(self.conf["cuda"]) for model in models.values()]
for model in models.values():
model["arguments"] = list(map(Variable, model["arguments"]))
model["labels"] = Variable(model["labels"])
# Train model
meters = {name: model_meter.ModelMeter(self.loader.distinct_labels) for name in models}
train_idx, val_idx = self.loader.train_idx, self.loader.val_idx
for epoch in range(self.conf["epochs"]):
for name, model_args in models.items():
self._train(epoch, name, model_args, train_idx, val_idx, meters[name])
# Testing
test_idx = self.loader.test_idx
for name, model_args in models.items():
meter = meters[name]
self._test(name, model_args, test_idx, meter)
self._data_logger.log_info(
model_name=name,
loss=meter.last_val("loss_test"),
acc=meter.last_val("acc_test"),
train_p=(train_p / (2 - train_p)) * 100,
norm_adj=self.conf["norm_adj"],
feat_type=self.conf["feat_type"]
)
# Currently supporting only binary class plotting
# meters[name].plot_auc(should_show=False)
# import matplotlib.pyplot as plt
# plt.savefig(os.path.join(self.products_path, time.strftime("%H_%M_%S_" + name)))
return meters
def _train(self, epoch, model_name, model_args, idx_train, idx_val, meter):
model, optimizer = model_args["model"], model_args["optimizer"]
arguments, labels = model_args["arguments"], model_args["labels"]
model.train()
optimizer.zero_grad()
output = model(*arguments)
loss_train = self._criterion(output[idx_train], labels[idx_train])
acc_train = model_meter.accuracy(output[idx_train], labels[idx_train])
meter.update_vals(loss_train=loss_train.item(), acc_train=acc_train)
loss_train.backward()
optimizer.step()
if not self.conf["fastmode"]:
# Evaluate validation set performance separately,
# deactivates dropout during validation run.
model.eval()
output = model(*arguments)
loss_val = self._criterion(output[idx_val], labels[idx_val])
acc_val = model_meter.accuracy(output[idx_val], labels[idx_val])
meter.update_vals(loss_val=loss_val.item(), acc_val=acc_val)
self._logger.debug("%s: Epoch: %03d, %s", model_name, epoch + 1, meter.log_str())
def _test(self, model_name, model_args, test_idx, meter):
model, arguments, labels = model_args["model"], model_args["arguments"], model_args["labels"]
model.eval()
output = model(*arguments)
loss_test = functional.nll_loss(output[test_idx], labels[test_idx])
acc_test = model_meter.accuracy(output[test_idx], labels[test_idx])
meter.update_diff(output[test_idx], labels[test_idx])
meter.update_vals(loss_test=loss_test.item(), acc_test=acc_test)
self._logger.info("%s: Test, %s", model_name, meter.log_str(log_vals=["loss_test", "acc_test"]))
# self._logger.info("%s Test: loss= %.4f accuracy= %.4f" % (model_name, loss_test.item(), acc_test.item()))
# return {"loss": loss_test.item(), "acc": acc_test.item()}
def init_seed(seed, cuda=None):
np.random.seed(seed)
torch.manual_seed(seed)
if cuda is not None:
torch.cuda.manual_seed(seed)
def aggregate_results(res_list, logger):
aggregated = {}
for cur_res in res_list:
for name, vals in cur_res.items():
if name not in aggregated:
aggregated[name] = {}
for key, val in vals.items():
if key not in aggregated[name]:
aggregated[name][key] = []
aggregated[name][key].append(val)
for name, vals in aggregated.items():
val_list = sorted(vals.items(), key=lambda x: x[0], reverse=True)
logger.info("*" * 15 + "%s mean: %s", name,
", ".join("%s=%3.4f" % (key, np.mean(val)) for key, val in val_list))
logger.info("*" * 15 + "%s std: %s", name, ", ".join("%s=%3.4f" % (key, np.std(val)) for key, val in val_list))
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument('--cuda', type=int, default=1,
help='Specify cuda device number')
parser.add_argument('--fastmode', action='store_true', default=False,
help='Validate during training pass.')
parser.add_argument('--seed', type=int, default=0, help='Random seed.')
parser.add_argument('--epochs', type=int, default=200,
help='Number of epochs to train.')
parser.add_argument('--dataset', type=str, default="cora",
help='The dataset to use.')
# parser.add_argument('--prefix', type=str, default="",
# help='The prefix of the products dir name.')
args = parser.parse_args()
# args.cuda = not args.no_cuda and torch.cuda.is_available()
if not torch.cuda.is_available():
args.cuda = None
return args
def main():
args = parse_args()
dataset = "cora" # args.dataset
seed = random.randint(1, 1000000000)
conf = {
"kipf": {"hidden": 16, "dropout": 0.5, "lr": 0.01, "weight_decay": 5e-4},
"hidden_layers": [16], "multi_hidden_layers": [100, 20], "dropout": 0.6, "lr": 0.01, "weight_decay": 0.001,
"dataset": dataset, "epochs": args.epochs, "cuda": args
|
ModelRunner
|
identifier_name
|
train.py
|
": [neighbor_features],
"features": [NODE_FEATURES],
"combined": [neighbor_features, NODE_FEATURES],
}
return dict(y for x in all_features[feat_type] for y in x.items())
class ModelRunner:
def __init__(self, products_path, dataset_path, conf, logger, data_logger=None):
self.conf = conf
self._logger = logger
self._data_logger = EmptyLogger() if data_logger is None else data_logger
self.products_path = products_path
self.loader = GraphLoader(dataset_path, is_max_connected=False, norm_adj=conf["norm_adj"],
cuda_num=conf["cuda"], logger=self._logger)
self._criterion = torch.nn.NLLLoss()
def _get_models(self):
bow_feat = self.loader.bow_mx
topo_feat = self.loader.topo_mx
model1 = GCN(nfeat=bow_feat.shape[1],
hlayers=[self.conf["kipf"]["hidden"]],
nclass=self.loader.num_labels,
dropout=self.conf["kipf"]["dropout"])
opt1 = optim.Adam(model1.parameters(), lr=self.conf["kipf"]["lr"],
weight_decay=self.conf["kipf"]["weight_decay"])
model2 = GCNCombined(nbow=bow_feat.shape[1],
nfeat=topo_feat.shape[1],
hlayers=self.conf["hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"])
opt2 = optim.Adam(model2.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model3 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=None)
opt3 = optim.Adam(model3.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model4 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=AsymmetricGCN)
opt4 = optim.Adam(model4.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
return {
"kipf": {
"model": model1, "optimizer": opt1,
"arguments": [self.loader.bow_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"our_combined": {
"model": model2, "optimizer": opt2,
"arguments": [self.loader.bow_mx, self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
"topo_sym": {
"model": model3, "optimizer": opt3,
"arguments": [self.loader.topo_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"topo_asym": {
"model": model4, "optimizer": opt4,
"arguments": [self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
}
def run(self, train_p, feat_type):
features_meta = get_features(feat_type, is_directed=self.loader.is_graph_directed)
self.loader.split_train(train_p, features_meta)
models = self._get_models()
if self.conf["cuda"] is not None:
[model["model"].cuda(self.conf["cuda"]) for model in models.values()]
for model in models.values():
model["arguments"] = list(map(Variable, model["arguments"]))
model["labels"] = Variable(model["labels"])
# Train model
meters = {name: model_meter.ModelMeter(self.loader.distinct_labels) for name in models}
train_idx, val_idx = self.loader.train_idx, self.loader.val_idx
for epoch in range(self.conf["epochs"]):
for name, model_args in models.items():
self._train(epoch, name, model_args, train_idx, val_idx, meters[name])
# Testing
test_idx = self.loader.test_idx
for name, model_args in models.items():
meter = meters[name]
self._test(name, model_args, test_idx, meter)
self._data_logger.log_info(
model_name=name,
loss=meter.last_val("loss_test"),
acc=meter.last_val("acc_test"),
train_p=(train_p / (2 - train_p)) * 100,
norm_adj=self.conf["norm_adj"],
feat_type=self.conf["feat_type"]
)
# Currently supporting only binary class plotting
# meters[name].plot_auc(should_show=False)
# import matplotlib.pyplot as plt
# plt.savefig(os.path.join(self.products_path, time.strftime("%H_%M_%S_" + name)))
return meters
def _train(self, epoch, model_name, model_args, idx_train, idx_val, meter):
model, optimizer = model_args["model"], model_args["optimizer"]
arguments, labels = model_args["arguments"], model_args["labels"]
model.train()
optimizer.zero_grad()
output = model(*arguments)
loss_train = self._criterion(output[idx_train], labels[idx_train])
acc_train = model_meter.accuracy(output[idx_train], labels[idx_train])
meter.update_vals(loss_train=loss_train.item(), acc_train=acc_train)
loss_train.backward()
optimizer.step()
if not self.conf["fastmode"]:
# Evaluate validation set performance separately,
# deactivates dropout during validation run.
model.eval()
output = model(*arguments)
loss_val = self._criterion(output[idx_val], labels[idx_val])
acc_val = model_meter.accuracy(output[idx_val], labels[idx_val])
meter.update_vals(loss_val=loss_val.item(), acc_val=acc_val)
self._logger.debug("%s: Epoch: %03d, %s", model_name, epoch + 1, meter.log_str())
def _test(self, model_name, model_args, test_idx, meter):
model, arguments, labels = model_args["model"], model_args["arguments"], model_args["labels"]
model.eval()
output = model(*arguments)
loss_test = functional.nll_loss(output[test_idx], labels[test_idx])
acc_test = model_meter.accuracy(output[test_idx], labels[test_idx])
meter.update_diff(output[test_idx], labels[test_idx])
meter.update_vals(loss_test=loss_test.item(), acc_test=acc_test)
self._logger.info("%s: Test, %s", model_name, meter.log_str(log_vals=["loss_test", "acc_test"]))
# self._logger.info("%s Test: loss= %.4f accuracy= %.4f" % (model_name, loss_test.item(), acc_test.item()))
# return {"loss": loss_test.item(), "acc": acc_test.item()}
def init_seed(seed, cuda=None):
np.random.seed(seed)
torch.manual_seed(seed)
if cuda is not None:
torch.cuda.manual_seed(seed)
def aggregate_results(res_list, logger):
aggregated = {}
for cur_res in res_list:
for name, vals in cur_res.items():
if name not in aggregated:
aggregated[name] = {}
for key, val in vals.items():
if key not in aggregated[name]:
aggregated[name][key] = []
aggregated[name][key].append(val)
for name, vals in aggregated.items():
|
val_list = sorted(vals.items(), key=lambda x: x[0], reverse=True)
logger.info("*" * 15 + "%s mean: %s", name,
", ".join("%s=%3.4f" % (key, np.mean(val)) for key, val in val_list))
logger.info("*" * 15 + "%s std: %s", name, ", ".join("%s=%3.4f" % (key, np.std(val)) for key, val in val_list))
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument('--cuda', type=int, default=1,
help='Specify cuda device number')
parser.add_argument('--fastmode', action='store_true', default=False,
help='Validate during training pass.')
parser.add_argument('--seed', type=int, default=0, help='Random seed.')
parser.add_argument('--epochs', type=int, default=200,
help='Number of epochs to train.')
parser.add_argument('--dataset', type=str, default="cora",
help='The dataset to use.')
# parser.add_argument('--prefix', type=str, default="",
# help='The prefix of the products dir name.')
args = parser.parse_args()
# args.cuda = not args.no_cuda and torch.cuda.is_available()
if not torch.cuda.is_available():
args.cuda = None
return args
def main():
args = parse_args()
dataset = "cora" # args.dataset
seed = random.randint(1, 1000000000)
conf = {
"kipf": {"hidden": 16, "dropout": 0.5, "lr": 0.01, "weight_decay": 5e-4},
"hidden_layers": [16], "multi_hidden_layers": [100, 20], "dropout": 0.6, "lr": 0.01, "weight_decay": 0.001,
"dataset": dataset, "epochs": args.epochs, "cuda": args.cuda
|
random_line_split
|
|
train.py
|
": [neighbor_features],
"features": [NODE_FEATURES],
"combined": [neighbor_features, NODE_FEATURES],
}
return dict(y for x in all_features[feat_type] for y in x.items())
class ModelRunner:
def __init__(self, products_path, dataset_path, conf, logger, data_logger=None):
self.conf = conf
self._logger = logger
self._data_logger = EmptyLogger() if data_logger is None else data_logger
self.products_path = products_path
self.loader = GraphLoader(dataset_path, is_max_connected=False, norm_adj=conf["norm_adj"],
cuda_num=conf["cuda"], logger=self._logger)
self._criterion = torch.nn.NLLLoss()
def _get_models(self):
bow_feat = self.loader.bow_mx
topo_feat = self.loader.topo_mx
model1 = GCN(nfeat=bow_feat.shape[1],
hlayers=[self.conf["kipf"]["hidden"]],
nclass=self.loader.num_labels,
dropout=self.conf["kipf"]["dropout"])
opt1 = optim.Adam(model1.parameters(), lr=self.conf["kipf"]["lr"],
weight_decay=self.conf["kipf"]["weight_decay"])
model2 = GCNCombined(nbow=bow_feat.shape[1],
nfeat=topo_feat.shape[1],
hlayers=self.conf["hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"])
opt2 = optim.Adam(model2.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model3 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=None)
opt3 = optim.Adam(model3.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model4 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=AsymmetricGCN)
opt4 = optim.Adam(model4.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
return {
"kipf": {
"model": model1, "optimizer": opt1,
"arguments": [self.loader.bow_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"our_combined": {
"model": model2, "optimizer": opt2,
"arguments": [self.loader.bow_mx, self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
"topo_sym": {
"model": model3, "optimizer": opt3,
"arguments": [self.loader.topo_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"topo_asym": {
"model": model4, "optimizer": opt4,
"arguments": [self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
}
def run(self, train_p, feat_type):
features_meta = get_features(feat_type, is_directed=self.loader.is_graph_directed)
self.loader.split_train(train_p, features_meta)
models = self._get_models()
if self.conf["cuda"] is not None:
[model["model"].cuda(self.conf["cuda"]) for model in models.values()]
for model in models.values():
model["arguments"] = list(map(Variable, model["arguments"]))
model["labels"] = Variable(model["labels"])
# Train model
meters = {name: model_meter.ModelMeter(self.loader.distinct_labels) for name in models}
train_idx, val_idx = self.loader.train_idx, self.loader.val_idx
for epoch in range(self.conf["epochs"]):
for name, model_args in models.items():
|
# Testing
test_idx = self.loader.test_idx
for name, model_args in models.items():
meter = meters[name]
self._test(name, model_args, test_idx, meter)
self._data_logger.log_info(
model_name=name,
loss=meter.last_val("loss_test"),
acc=meter.last_val("acc_test"),
train_p=(train_p / (2 - train_p)) * 100,
norm_adj=self.conf["norm_adj"],
feat_type=self.conf["feat_type"]
)
# Currently supporting only binary class plotting
# meters[name].plot_auc(should_show=False)
# import matplotlib.pyplot as plt
# plt.savefig(os.path.join(self.products_path, time.strftime("%H_%M_%S_" + name)))
return meters
def _train(self, epoch, model_name, model_args, idx_train, idx_val, meter):
model, optimizer = model_args["model"], model_args["optimizer"]
arguments, labels = model_args["arguments"], model_args["labels"]
model.train()
optimizer.zero_grad()
output = model(*arguments)
loss_train = self._criterion(output[idx_train], labels[idx_train])
acc_train = model_meter.accuracy(output[idx_train], labels[idx_train])
meter.update_vals(loss_train=loss_train.item(), acc_train=acc_train)
loss_train.backward()
optimizer.step()
if not self.conf["fastmode"]:
# Evaluate validation set performance separately,
# deactivates dropout during validation run.
model.eval()
output = model(*arguments)
loss_val = self._criterion(output[idx_val], labels[idx_val])
acc_val = model_meter.accuracy(output[idx_val], labels[idx_val])
meter.update_vals(loss_val=loss_val.item(), acc_val=acc_val)
self._logger.debug("%s: Epoch: %03d, %s", model_name, epoch + 1, meter.log_str())
def _test(self, model_name, model_args, test_idx, meter):
model, arguments, labels = model_args["model"], model_args["arguments"], model_args["labels"]
model.eval()
output = model(*arguments)
loss_test = functional.nll_loss(output[test_idx], labels[test_idx])
acc_test = model_meter.accuracy(output[test_idx], labels[test_idx])
meter.update_diff(output[test_idx], labels[test_idx])
meter.update_vals(loss_test=loss_test.item(), acc_test=acc_test)
self._logger.info("%s: Test, %s", model_name, meter.log_str(log_vals=["loss_test", "acc_test"]))
# self._logger.info("%s Test: loss= %.4f accuracy= %.4f" % (model_name, loss_test.item(), acc_test.item()))
# return {"loss": loss_test.item(), "acc": acc_test.item()}
def init_seed(seed, cuda=None):
np.random.seed(seed)
torch.manual_seed(seed)
if cuda is not None:
torch.cuda.manual_seed(seed)
def aggregate_results(res_list, logger):
aggregated = {}
for cur_res in res_list:
for name, vals in cur_res.items():
if name not in aggregated:
aggregated[name] = {}
for key, val in vals.items():
if key not in aggregated[name]:
aggregated[name][key] = []
aggregated[name][key].append(val)
for name, vals in aggregated.items():
val_list = sorted(vals.items(), key=lambda x: x[0], reverse=True)
logger.info("*" * 15 + "%s mean: %s", name,
", ".join("%s=%3.4f" % (key, np.mean(val)) for key, val in val_list))
logger.info("*" * 15 + "%s std: %s", name, ", ".join("%s=%3.4f" % (key, np.std(val)) for key, val in val_list))
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument('--cuda', type=int, default=1,
help='Specify cuda device number')
parser.add_argument('--fastmode', action='store_true', default=False,
help='Validate during training pass.')
parser.add_argument('--seed', type=int, default=0, help='Random seed.')
parser.add_argument('--epochs', type=int, default=200,
help='Number of epochs to train.')
parser.add_argument('--dataset', type=str, default="cora",
help='The dataset to use.')
# parser.add_argument('--prefix', type=str, default="",
# help='The prefix of the products dir name.')
args = parser.parse_args()
# args.cuda = not args.no_cuda and torch.cuda.is_available()
if not torch.cuda.is_available():
args.cuda = None
return args
def main():
args = parse_args()
dataset = "cora" # args.dataset
seed = random.randint(1, 1000000000)
conf = {
"kipf": {"hidden": 16, "dropout": 0.5, "lr": 0.01, "weight_decay": 5e-4},
"hidden_layers": [16], "multi_hidden_layers": [100, 20], "dropout": 0.6, "lr": 0.01, "weight_decay": 0.001,
"dataset": dataset, "epochs": args.epochs, "cuda": args
|
self._train(epoch, name, model_args, train_idx, val_idx, meters[name])
|
conditional_block
|
train.py
|
": [neighbor_features],
"features": [NODE_FEATURES],
"combined": [neighbor_features, NODE_FEATURES],
}
return dict(y for x in all_features[feat_type] for y in x.items())
class ModelRunner:
def __init__(self, products_path, dataset_path, conf, logger, data_logger=None):
self.conf = conf
self._logger = logger
self._data_logger = EmptyLogger() if data_logger is None else data_logger
self.products_path = products_path
self.loader = GraphLoader(dataset_path, is_max_connected=False, norm_adj=conf["norm_adj"],
cuda_num=conf["cuda"], logger=self._logger)
self._criterion = torch.nn.NLLLoss()
def _get_models(self):
|
dropout=self.conf["dropout"],
layer_type=None)
opt3 = optim.Adam(model3.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model4 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"],
layer_type=AsymmetricGCN)
opt4 = optim.Adam(model4.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
return {
"kipf": {
"model": model1, "optimizer": opt1,
"arguments": [self.loader.bow_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"our_combined": {
"model": model2, "optimizer": opt2,
"arguments": [self.loader.bow_mx, self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
"topo_sym": {
"model": model3, "optimizer": opt3,
"arguments": [self.loader.topo_mx, self.loader.adj_mx],
"labels": self.loader.labels,
},
"topo_asym": {
"model": model4, "optimizer": opt4,
"arguments": [self.loader.topo_mx, self.loader.adj_rt_mx],
"labels": self.loader.labels,
},
}
def run(self, train_p, feat_type):
features_meta = get_features(feat_type, is_directed=self.loader.is_graph_directed)
self.loader.split_train(train_p, features_meta)
models = self._get_models()
if self.conf["cuda"] is not None:
[model["model"].cuda(self.conf["cuda"]) for model in models.values()]
for model in models.values():
model["arguments"] = list(map(Variable, model["arguments"]))
model["labels"] = Variable(model["labels"])
# Train model
meters = {name: model_meter.ModelMeter(self.loader.distinct_labels) for name in models}
train_idx, val_idx = self.loader.train_idx, self.loader.val_idx
for epoch in range(self.conf["epochs"]):
for name, model_args in models.items():
self._train(epoch, name, model_args, train_idx, val_idx, meters[name])
# Testing
test_idx = self.loader.test_idx
for name, model_args in models.items():
meter = meters[name]
self._test(name, model_args, test_idx, meter)
self._data_logger.log_info(
model_name=name,
loss=meter.last_val("loss_test"),
acc=meter.last_val("acc_test"),
train_p=(train_p / (2 - train_p)) * 100,
norm_adj=self.conf["norm_adj"],
feat_type=self.conf["feat_type"]
)
# Currently supporting only binary class plotting
# meters[name].plot_auc(should_show=False)
# import matplotlib.pyplot as plt
# plt.savefig(os.path.join(self.products_path, time.strftime("%H_%M_%S_" + name)))
return meters
def _train(self, epoch, model_name, model_args, idx_train, idx_val, meter):
model, optimizer = model_args["model"], model_args["optimizer"]
arguments, labels = model_args["arguments"], model_args["labels"]
model.train()
optimizer.zero_grad()
output = model(*arguments)
loss_train = self._criterion(output[idx_train], labels[idx_train])
acc_train = model_meter.accuracy(output[idx_train], labels[idx_train])
meter.update_vals(loss_train=loss_train.item(), acc_train=acc_train)
loss_train.backward()
optimizer.step()
if not self.conf["fastmode"]:
# Evaluate validation set performance separately,
# deactivates dropout during validation run.
model.eval()
output = model(*arguments)
loss_val = self._criterion(output[idx_val], labels[idx_val])
acc_val = model_meter.accuracy(output[idx_val], labels[idx_val])
meter.update_vals(loss_val=loss_val.item(), acc_val=acc_val)
self._logger.debug("%s: Epoch: %03d, %s", model_name, epoch + 1, meter.log_str())
def _test(self, model_name, model_args, test_idx, meter):
model, arguments, labels = model_args["model"], model_args["arguments"], model_args["labels"]
model.eval()
output = model(*arguments)
loss_test = functional.nll_loss(output[test_idx], labels[test_idx])
acc_test = model_meter.accuracy(output[test_idx], labels[test_idx])
meter.update_diff(output[test_idx], labels[test_idx])
meter.update_vals(loss_test=loss_test.item(), acc_test=acc_test)
self._logger.info("%s: Test, %s", model_name, meter.log_str(log_vals=["loss_test", "acc_test"]))
# self._logger.info("%s Test: loss= %.4f accuracy= %.4f" % (model_name, loss_test.item(), acc_test.item()))
# return {"loss": loss_test.item(), "acc": acc_test.item()}
def init_seed(seed, cuda=None):
np.random.seed(seed)
torch.manual_seed(seed)
if cuda is not None:
torch.cuda.manual_seed(seed)
def aggregate_results(res_list, logger):
aggregated = {}
for cur_res in res_list:
for name, vals in cur_res.items():
if name not in aggregated:
aggregated[name] = {}
for key, val in vals.items():
if key not in aggregated[name]:
aggregated[name][key] = []
aggregated[name][key].append(val)
for name, vals in aggregated.items():
val_list = sorted(vals.items(), key=lambda x: x[0], reverse=True)
logger.info("*" * 15 + "%s mean: %s", name,
", ".join("%s=%3.4f" % (key, np.mean(val)) for key, val in val_list))
logger.info("*" * 15 + "%s std: %s", name, ", ".join("%s=%3.4f" % (key, np.std(val)) for key, val in val_list))
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument('--cuda', type=int, default=1,
help='Specify cuda device number')
parser.add_argument('--fastmode', action='store_true', default=False,
help='Validate during training pass.')
parser.add_argument('--seed', type=int, default=0, help='Random seed.')
parser.add_argument('--epochs', type=int, default=200,
help='Number of epochs to train.')
parser.add_argument('--dataset', type=str, default="cora",
help='The dataset to use.')
# parser.add_argument('--prefix', type=str, default="",
# help='The prefix of the products dir name.')
args = parser.parse_args()
# args.cuda = not args.no_cuda and torch.cuda.is_available()
if not torch.cuda.is_available():
args.cuda = None
return args
def main():
args = parse_args()
dataset = "cora" # args.dataset
seed = random.randint(1, 1000000000)
conf = {
"kipf": {"hidden": 16, "dropout": 0.5, "lr": 0.01, "weight_decay": 5e-4},
"hidden_layers": [16], "multi_hidden_layers": [100, 20], "dropout": 0.6, "lr": 0.01, "weight_decay": 0.001,
"dataset": dataset, "epochs": args.epochs, "cuda": args.cuda
|
bow_feat = self.loader.bow_mx
topo_feat = self.loader.topo_mx
model1 = GCN(nfeat=bow_feat.shape[1],
hlayers=[self.conf["kipf"]["hidden"]],
nclass=self.loader.num_labels,
dropout=self.conf["kipf"]["dropout"])
opt1 = optim.Adam(model1.parameters(), lr=self.conf["kipf"]["lr"],
weight_decay=self.conf["kipf"]["weight_decay"])
model2 = GCNCombined(nbow=bow_feat.shape[1],
nfeat=topo_feat.shape[1],
hlayers=self.conf["hidden_layers"],
nclass=self.loader.num_labels,
dropout=self.conf["dropout"])
opt2 = optim.Adam(model2.parameters(), lr=self.conf["lr"], weight_decay=self.conf["weight_decay"])
model3 = GCN(nfeat=topo_feat.shape[1],
hlayers=self.conf["multi_hidden_layers"],
nclass=self.loader.num_labels,
|
identifier_body
|
CAAPR_Pipeline.py
|
_async( CAAPR.CAAPR_Photom.SubpipelinePhotom, args=(source_dict, bands_dict[band], kwargs_dict) ) )
pool.close()
pool.join()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Gathering parallel threads.'
photom_output_list = [output.get() for output in photom_output_list if output.successful()==True]
photom_list = [photom for photom in photom_output_list if photom!=None]
# If parallelisation is disabled, process sources one-at-a-time
elif kwargs_dict['parallel']==False:
for band in bands_dict.keys():
photom_output_list.append( CAAPR.CAAPR_Photom.SubpipelinePhotom(source_dict, bands_dict[band], kwargs_dict) )
photom_list = [photom for photom in photom_output_list if photom!=None]
# Shortcut if no photometry is done
if not kwargs_dict['do_photom']:
photom_attempts = 'Complete'
gc.collect()
return
# Check that all photometry completed
photom_attempts, photom_output_list = CAAPR.CAAPR_Photom.PhotomCheck(photom_attempts, photom_output_list, source_dict, bands_dict, kwargs_dict)
# Record photometry results to file
CAAPR.CAAPR_IO.RecordPhotom(photom_list, source_dict, bands_dict, kwargs_dict)
# Prepare thumbnail images for bands excluded from photometry
CAAPR.CAAPR_Photom.ExcludedThumb(source_dict, bands_dict, kwargs_dict)
# Create grid of thumbnail images
CAAPR.CAAPR_IO.PhotomThumbGrid(source_dict, bands_dict, kwargs_dict)
# Report time taken to do photometry, and tidy up
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Time taken performing actual photometry: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-photom_start,4))+' seconds.'
# Tidy up temporary files and paths
bands_dict = PathTidy(source_dict, bands_dict, kwargs_dict)
# Report time taken for source, and tidy up garbage
gc.collect()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Total time taken for souce: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-source_start,4))+' seconds.'
if kwargs_dict['thumbnails']==True and kwargs_dict['messy']==False:
[os.remove(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps',processed_map)) for processed_map in os.listdir(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps')) if '.fits' in processed_map]
# Define function to check if data actually exists for any band for this source
def SourcePrelim(source_dict, bands_dict, kwargs_dict):
# Check that any of the bands actually have data for this source
kwargs_dict_copy = copy.deepcopy(kwargs_dict)
kwargs_dict_copy['verbose'] = False
bands_check = []
for band in bands_dict.keys():
source_id = source_dict['name']+'_'+bands_dict[band]['band_name']
in_fitspath, file_found = CAAPR.CAAPR_Pipeline.FilePrelim(source_dict, bands_dict[band], kwargs_dict_copy)
bands_check.append(file_found)
# Report to user if no data found
if True not in bands_check:
print '['+source_id+'] No data found in target directory for current source.'
# Make null entries in tables, as necessary
if kwargs_dict['fit_apertures']==True:
null_aperture_combined = [np.NaN, np.NaN, np.NaN, np.NaN]
CAAPR.CAAPR_IO.RecordAperture(null_aperture_combined, source_dict, kwargs_dict)
if kwargs_dict['do_photom']==True:
CAAPR.CAAPR_IO.RecordPhotom([], source_dict, bands_dict, kwargs_dict)
# Return result
if True not in bands_check:
return False
elif True in bands_check:
return True
# Define function that does basic initial handling of band parameters
def BandInitiate(band_dict):
# Make sure band has content
if band_dict==None:
return band_dict
# Parse band cutout request, converting string to boolean if necessary
if band_dict['make_cutout']=='True':
band_dict['make_cutout']=True
elif band_dict['make_cutout']=='False':
band_dict['make_cutout']=False
else:
try:
band_dict['make_cutout'] = float(band_dict['make_cutout'])
except:
raise Exception('Cutout request not understood; should either be False, or width of cutout in arcseconds.')
# Reset band directory to inviolate value, to purge any holdovers from previous source
band_dict['band_dir'] = band_dict['band_dir_inviolate']
# Return band dict
return band_dict
# Define function that performs preimilary checks of file type and location
def FilePrelim(source_dict, band_dict, kwargs_dict):
# Determine whether the user is specificing a directroy full of FITS files in this band (in which case use standardised filename format), or just a single FITS file
try:
if os.path.isdir(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'], source_dict['name']+'_'+band_dict['band_name'] )
elif os.path.isfile(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'] )
except:
pdb.set_trace()
# Work out whether the file extension for FITS file in question is .fits or .fits.gz
file_found = False
try:
if os.path.exists(in_fitspath+'.fits'):
in_fitspath = in_fitspath+'.fits'
file_found = True
elif os.path.exists(in_fitspath+'.fits.gz'):
in_fitspath = in_fitspath+'.fits.gz'
file_found = True
except:
raise Exception('Path provided for band '+str(band_dict['band_name'])+' refers to neither a file nor a folder.')
# Return file values
return in_fitspath, file_found
# Initiate the pod (Photometry Organisation Dictionary)
def PodInitiate(in_fitspath, source_dict, band_dict, kwargs_dict):
source_id = source_dict['name']+'_'+band_dict['band_name']
if kwargs_dict['verbose']: print '['+source_id+'] Reading in FITS data.'
# Read in FITS file in question
in_fitsdata = astropy.io.fits.open(in_fitspath)
in_image = in_fitsdata[0].data
in_header = in_fitsdata[0].header
in_fitsdata.close()
in_wcs = astropy.wcs.WCS(in_header)
in_fitspath_size = float(os.stat(in_fitspath).st_size)
# Create the pod (Photometry Organisation Dictionary), which will bundle all the photometry data for this source & band into one dictionary to be passed between functions
pod = {'in_fitspath':in_fitspath,
'in_image':in_image,
'in_header':in_header,
'in_wcs':in_wcs,
'cutout':in_image.copy(),
'output_dir_path':kwargs_dict['output_dir_path'],
'temp_dir_path':kwargs_dict['temp_dir_path'],
'in_fitspath_size':in_fitspath_size,
'id':source_id,
'verbose':kwargs_dict['verbose']}
# Return pod
return pod
# Define function that determines preliminary map values
def MapPrelim(pod, source_dict, band_dict, verbose=False):
|
if pod['verbose']: print '['+pod['id']+'] Determining properties of map.'
# Check if x & y pixel sizes are meaningfully different. If so, panic; else, treat as same
pix_size = 3600.0 * pod['in_wcs'].wcs.cdelt
if float(abs(pix_size.max()))/float(abs(pix_size.min()))>(1+1E-3):
raise Exception('The x pixel size if noticably different from the y pixel size.')
else:
pod['pix_arcsec'] = float(np.mean(np.abs(pix_size)))
# Determine source position in cutout in ij coordinates, and size of cutout
centre_xy = pod['in_wcs'].wcs_world2pix( np.array([[ source_dict['ra'], source_dict['dec'] ]]), 0 )
pod['centre_i'], pod['centre_j'] = float(centre_xy[0][1]), float(centre_xy[0][0])
pod['box_rad'] = int( round( float(pod['cutout'].shape[0]) * 0.5 ) )
# Determine beam size in pixels; if beam size not given, then assume map is Nyquist sampled (ie, 2.355 pixels ber beam)
if isinstance(band_dict['beam_arcsec'], numbers.Number):
pod['beam_pix'] = float(band_dict['beam_arcsec']) / pod['pix_arcsec']
else:
|
identifier_body
|
|
CAAPR_Pipeline.py
|
, source_dict, kwargs_dict)
# Record aperture properties to file
CAAPR.CAAPR_IO.RecordAperture(aperture_combined, source_dict, kwargs_dict)
# Prepare thumbnail images for bands excluded from aperture fitting
CAAPR.CAAPR_Aperture.ExcludedThumb(source_dict, bands_dict, kwargs_dict, aperture_list, aperture_combined)
# Create grid of thumbnail images
CAAPR.CAAPR_IO.ApertureThumbGrid(source_dict, bands_dict, kwargs_dict, aperture_list, aperture_combined)
# Report time taken to fit apertures, and tidy up garbage
gc.collect()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Time taken performing aperture fitting: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-aperture_start,4))+' seconds.'
# Commence actual photometry sub-pipeline
if kwargs_dict['do_photom'] or kwargs_dict['save_images']:
# Handle problem where the user hasn't provided an aperture file, but also hasn't told CAAPR to fit its own apertures.
if kwargs_dict['aperture_table_path']==False and kwargs_dict['fit_apertures']==False:
raise Exception('User has requested no aperture-fitting, and no photometry!')
# Process sources inside while loop, to catch 'missed' bands
photom_attempts = 0
while photom_attempts!='Complete':
# In standard operation, process multiple sources in parallel
photom_start = time.time()
photom_output_list = []
if kwargs_dict['parallel']==True:
bands_dict_keys = bands_dict.keys()
random.shuffle(bands_dict_keys)
pool = mp.Pool(processes=kwargs_dict['n_proc'])
for band in bands_dict_keys:
photom_output_list.append( pool.apply_async( CAAPR.CAAPR_Photom.SubpipelinePhotom, args=(source_dict, bands_dict[band], kwargs_dict) ) )
pool.close()
pool.join()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Gathering parallel threads.'
photom_output_list = [output.get() for output in photom_output_list if output.successful()==True]
photom_list = [photom for photom in photom_output_list if photom!=None]
# If parallelisation is disabled, process sources one-at-a-time
elif kwargs_dict['parallel']==False:
for band in bands_dict.keys():
photom_output_list.append( CAAPR.CAAPR_Photom.SubpipelinePhotom(source_dict, bands_dict[band], kwargs_dict) )
photom_list = [photom for photom in photom_output_list if photom!=None]
# Shortcut if no photometry is done
if not kwargs_dict['do_photom']:
photom_attempts = 'Complete'
gc.collect()
return
# Check that all photometry completed
photom_attempts, photom_output_list = CAAPR.CAAPR_Photom.PhotomCheck(photom_attempts, photom_output_list, source_dict, bands_dict, kwargs_dict)
# Record photometry results to file
CAAPR.CAAPR_IO.RecordPhotom(photom_list, source_dict, bands_dict, kwargs_dict)
# Prepare thumbnail images for bands excluded from photometry
CAAPR.CAAPR_Photom.ExcludedThumb(source_dict, bands_dict, kwargs_dict)
# Create grid of thumbnail images
CAAPR.CAAPR_IO.PhotomThumbGrid(source_dict, bands_dict, kwargs_dict)
# Report time taken to do photometry, and tidy up
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Time taken performing actual photometry: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-photom_start,4))+' seconds.'
# Tidy up temporary files and paths
bands_dict = PathTidy(source_dict, bands_dict, kwargs_dict)
# Report time taken for source, and tidy up garbage
gc.collect()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Total time taken for souce: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-source_start,4))+' seconds.'
if kwargs_dict['thumbnails']==True and kwargs_dict['messy']==False:
[os.remove(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps',processed_map)) for processed_map in os.listdir(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps')) if '.fits' in processed_map]
# Define function to check if data actually exists for any band for this source
def SourcePrelim(source_dict, bands_dict, kwargs_dict):
# Check that any of the bands actually have data for this source
kwargs_dict_copy = copy.deepcopy(kwargs_dict)
kwargs_dict_copy['verbose'] = False
bands_check = []
for band in bands_dict.keys():
source_id = source_dict['name']+'_'+bands_dict[band]['band_name']
in_fitspath, file_found = CAAPR.CAAPR_Pipeline.FilePrelim(source_dict, bands_dict[band], kwargs_dict_copy)
bands_check.append(file_found)
# Report to user if no data found
if True not in bands_check:
print '['+source_id+'] No data found in target directory for current source.'
# Make null entries in tables, as necessary
if kwargs_dict['fit_apertures']==True:
null_aperture_combined = [np.NaN, np.NaN, np.NaN, np.NaN]
CAAPR.CAAPR_IO.RecordAperture(null_aperture_combined, source_dict, kwargs_dict)
if kwargs_dict['do_photom']==True:
CAAPR.CAAPR_IO.RecordPhotom([], source_dict, bands_dict, kwargs_dict)
# Return result
if True not in bands_check:
return False
elif True in bands_check:
return True
# Define function that does basic initial handling of band parameters
def BandInitiate(band_dict):
# Make sure band has content
if band_dict==None:
return band_dict
# Parse band cutout request, converting string to boolean if necessary
if band_dict['make_cutout']=='True':
band_dict['make_cutout']=True
elif band_dict['make_cutout']=='False':
band_dict['make_cutout']=False
else:
try:
band_dict['make_cutout'] = float(band_dict['make_cutout'])
except:
raise Exception('Cutout request not understood; should either be False, or width of cutout in arcseconds.')
# Reset band directory to inviolate value, to purge any holdovers from previous source
band_dict['band_dir'] = band_dict['band_dir_inviolate']
# Return band dict
return band_dict
# Define function that performs preimilary checks of file type and location
def
|
(source_dict, band_dict, kwargs_dict):
# Determine whether the user is specificing a directroy full of FITS files in this band (in which case use standardised filename format), or just a single FITS file
try:
if os.path.isdir(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'], source_dict['name']+'_'+band_dict['band_name'] )
elif os.path.isfile(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'] )
except:
pdb.set_trace()
# Work out whether the file extension for FITS file in question is .fits or .fits.gz
file_found = False
try:
if os.path.exists(in_fitspath+'.fits'):
in_fitspath = in_fitspath+'.fits'
file_found = True
elif os.path.exists(in_fitspath+'.fits.gz'):
in_fitspath = in_fitspath+'.fits.gz'
file_found = True
except:
raise Exception('Path provided for band '+str(band_dict['band_name'])+' refers to neither a file nor a folder.')
# Return file values
return in_fitspath, file_found
# Initiate the pod (Photometry Organisation Dictionary)
def PodInitiate(in_fitspath, source_dict, band_dict, kwargs_dict):
source_id = source_dict['name']+'_'+band_dict['band_name']
if kwargs_dict['verbose']: print '['+source_id+'] Reading in FITS data.'
# Read in FITS file in question
in_fitsdata = astropy.io.fits.open(in_fitspath)
in_image = in_fitsdata[0].data
in_header = in_fitsdata[0].header
in_fitsdata.close()
in_wcs = astropy.wcs.WCS(in_header)
in_fitspath_size = float(os.stat(in_fitspath).st_size)
# Create the pod (Photometry Organisation Dictionary), which will bundle all the photometry data for this source & band into one dictionary to be passed between functions
pod = {'in_fitspath':in_fitspath,
'in_image':in_image,
'in_header':in_header,
'in_wcs':in_wcs,
'cutout':in_image.copy(),
'output_dir_path':kwargs_dict['output_dir_path'],
'temp
|
FilePrelim
|
identifier_name
|
CAAPR_Pipeline.py
|
format), or just a single FITS file
try:
if os.path.isdir(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'], source_dict['name']+'_'+band_dict['band_name'] )
elif os.path.isfile(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'] )
except:
pdb.set_trace()
# Work out whether the file extension for FITS file in question is .fits or .fits.gz
file_found = False
try:
if os.path.exists(in_fitspath+'.fits'):
in_fitspath = in_fitspath+'.fits'
file_found = True
elif os.path.exists(in_fitspath+'.fits.gz'):
in_fitspath = in_fitspath+'.fits.gz'
file_found = True
except:
raise Exception('Path provided for band '+str(band_dict['band_name'])+' refers to neither a file nor a folder.')
# Return file values
return in_fitspath, file_found
# Initiate the pod (Photometry Organisation Dictionary)
def PodInitiate(in_fitspath, source_dict, band_dict, kwargs_dict):
source_id = source_dict['name']+'_'+band_dict['band_name']
if kwargs_dict['verbose']: print '['+source_id+'] Reading in FITS data.'
# Read in FITS file in question
in_fitsdata = astropy.io.fits.open(in_fitspath)
in_image = in_fitsdata[0].data
in_header = in_fitsdata[0].header
in_fitsdata.close()
in_wcs = astropy.wcs.WCS(in_header)
in_fitspath_size = float(os.stat(in_fitspath).st_size)
# Create the pod (Photometry Organisation Dictionary), which will bundle all the photometry data for this source & band into one dictionary to be passed between functions
pod = {'in_fitspath':in_fitspath,
'in_image':in_image,
'in_header':in_header,
'in_wcs':in_wcs,
'cutout':in_image.copy(),
'output_dir_path':kwargs_dict['output_dir_path'],
'temp_dir_path':kwargs_dict['temp_dir_path'],
'in_fitspath_size':in_fitspath_size,
'id':source_id,
'verbose':kwargs_dict['verbose']}
# Return pod
return pod
# Define function that determines preliminary map values
def MapPrelim(pod, source_dict, band_dict, verbose=False):
if pod['verbose']: print '['+pod['id']+'] Determining properties of map.'
# Check if x & y pixel sizes are meaningfully different. If so, panic; else, treat as same
pix_size = 3600.0 * pod['in_wcs'].wcs.cdelt
if float(abs(pix_size.max()))/float(abs(pix_size.min()))>(1+1E-3):
raise Exception('The x pixel size if noticably different from the y pixel size.')
else:
pod['pix_arcsec'] = float(np.mean(np.abs(pix_size)))
# Determine source position in cutout in ij coordinates, and size of cutout
centre_xy = pod['in_wcs'].wcs_world2pix( np.array([[ source_dict['ra'], source_dict['dec'] ]]), 0 )
pod['centre_i'], pod['centre_j'] = float(centre_xy[0][1]), float(centre_xy[0][0])
pod['box_rad'] = int( round( float(pod['cutout'].shape[0]) * 0.5 ) )
# Determine beam size in pixels; if beam size not given, then assume map is Nyquist sampled (ie, 2.355 pixels ber beam)
if isinstance(band_dict['beam_arcsec'], numbers.Number):
pod['beam_pix'] = float(band_dict['beam_arcsec']) / pod['pix_arcsec']
else:
pod['beam_pix'] = pod['pix_arcsec'] * 2.355
# Check if current source lies within bounds of map; if not, fai and return)
if pod['centre_i']<0 or pod['centre_i']>(pod['cutout'].shape)[0] or pod['centre_j']<0 or pod['centre_j']>(pod['cutout'].shape)[1]:
pod['within_bounds'] = False
if 'band_dir_inviolate' in band_dict.keys():
band_dict['band_dir'] = band_dict['band_dir_inviolate']
if pod['verbose']: print '['+pod['id']+'] Target not within bounds of map.'
else:
pod['within_bounds'] = True
# Return pod
return pod
# Define function that fits and subtracts polynomial background filter from map
def PolySub(pod, mask_semimaj_pix, mask_axial_ratio, mask_angle, poly_order=5, cutoff_sigma=2.0, instant_quit=False):
if pod['verbose']: print '['+pod['id']+'] Determining if (and how) background is significnatly variable.'
# If polynomial background subraction not wanted, immediately return everything unchanged
if instant_quit:
pod['sky_poly'] = False
return pod
# If image has pixels smaller than some limit, downsample image to improve processing time
pix_size = pod['pix_arcsec']
pix_size_limit = 2.0
if pix_size<pix_size_limit:
downsample_factor = int(np.ceil(pix_size_limit/pix_size))
else:
downsample_factor = 1
image_ds = ChrisFuncs.Downsample(pod['cutout'], downsample_factor)
# Downsample related values accordingly
mask_semimaj_pix = mask_semimaj_pix / downsample_factor
centre_i = int(round(float((0.5*pod['centre_i'])-1.0)))
centre_j = int(round(float((0.5*pod['centre_j'])-1.0)))
# Find cutoff for excluding bright pixels by sigma-clipping map
clip_value = ChrisFuncs.SigmaClip(image_ds, tolerance=0.01, sigma_thresh=3.0, median=True)
noise_value = clip_value[0]
field_value = clip_value[1]
cutoff = field_value + ( cutoff_sigma * noise_value )
# Mask all image pixels in masking region around source
image_masked = image_ds.copy()
ellipse_mask = ChrisFuncs.Photom.EllipseMask(image_ds, mask_semimaj_pix, mask_axial_ratio, mask_angle, centre_i, centre_j)
image_masked[ np.where( ellipse_mask==1 ) ] = np.nan
# Mask all image pixels identified as being high SNR
image_masked[ np.where( image_masked>cutoff ) ] = np.nan
# Use astropy to set up 2-dimensional polynomial to the image
image_masked[ np.where( np.isnan(image_masked)==True ) ] = field_value
poly_model = astropy.modeling.models.Polynomial2D(degree=poly_order)
i_coords, j_coords = np.mgrid[:image_masked.shape[0], :image_masked.shape[1]]
fitter = astropy.modeling.fitting.LevMarLSQFitter()
i_coords = i_coords.flatten()
j_coords = j_coords.flatten()
image_flattened = image_masked.flatten()
good = np.where(np.isnan(image_flattened)==False)
i_coords = i_coords[good]
j_coords = j_coords[good]
# Attempt polynomial fit; if insufficient data then skip onwards
image_flattened = image_flattened[good]
try:
fit = fitter(poly_model, i_coords, j_coords, image_flattened)
except:
if pod['verbose']: print '['+pod['id']+'] Background is not significnatly variable; leaving image unaltered.'
pod['sky_poly'] = False
return pod
# Create final polynomial filter (undoing downsampling using lorenzoriano GitHub script)
i_coords, j_coords = np.mgrid[:image_ds.shape[0], :image_ds.shape[1]]
poly_fit = fit(i_coords, j_coords)
poly_full = scipy.ndimage.interpolation.zoom(poly_fit, [ float(pod['cutout'].shape[0])/float(poly_fit.shape[0]), float(pod['cutout'].shape[1])/float(poly_fit.shape[1]) ], mode='nearest') #poly_full = congrid.congrid(poly_fit, (pod['cutout'].shape[0], pod['cutout'].shape[1]), minusone=True)
# Establish background variation before application of filter
sigma_thresh = 3.0
clip_in = ChrisFuncs.SigmaClip(pod['cutout'], tolerance=0.005, median=True, sigma_thresh=sigma_thresh)
bg_in = pod['cutout'][ np.where( pod['cutout']<clip_in[1] ) ]
|
spread_in = np.mean( np.abs( bg_in - clip_in[1] ) )
# How much reduction in background variation there was due to application of the filter
image_sub = pod['cutout'] - poly_full
|
random_line_split
|
|
CAAPR_Pipeline.py
|
, source_dict, kwargs_dict)
# Record aperture properties to file
CAAPR.CAAPR_IO.RecordAperture(aperture_combined, source_dict, kwargs_dict)
# Prepare thumbnail images for bands excluded from aperture fitting
CAAPR.CAAPR_Aperture.ExcludedThumb(source_dict, bands_dict, kwargs_dict, aperture_list, aperture_combined)
# Create grid of thumbnail images
CAAPR.CAAPR_IO.ApertureThumbGrid(source_dict, bands_dict, kwargs_dict, aperture_list, aperture_combined)
# Report time taken to fit apertures, and tidy up garbage
gc.collect()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Time taken performing aperture fitting: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-aperture_start,4))+' seconds.'
# Commence actual photometry sub-pipeline
if kwargs_dict['do_photom'] or kwargs_dict['save_images']:
# Handle problem where the user hasn't provided an aperture file, but also hasn't told CAAPR to fit its own apertures.
if kwargs_dict['aperture_table_path']==False and kwargs_dict['fit_apertures']==False:
raise Exception('User has requested no aperture-fitting, and no photometry!')
# Process sources inside while loop, to catch 'missed' bands
photom_attempts = 0
while photom_attempts!='Complete':
# In standard operation, process multiple sources in parallel
photom_start = time.time()
photom_output_list = []
if kwargs_dict['parallel']==True:
bands_dict_keys = bands_dict.keys()
random.shuffle(bands_dict_keys)
pool = mp.Pool(processes=kwargs_dict['n_proc'])
for band in bands_dict_keys:
photom_output_list.append( pool.apply_async( CAAPR.CAAPR_Photom.SubpipelinePhotom, args=(source_dict, bands_dict[band], kwargs_dict) ) )
pool.close()
pool.join()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Gathering parallel threads.'
photom_output_list = [output.get() for output in photom_output_list if output.successful()==True]
photom_list = [photom for photom in photom_output_list if photom!=None]
# If parallelisation is disabled, process sources one-at-a-time
elif kwargs_dict['parallel']==False:
for band in bands_dict.keys():
photom_output_list.append( CAAPR.CAAPR_Photom.SubpipelinePhotom(source_dict, bands_dict[band], kwargs_dict) )
photom_list = [photom for photom in photom_output_list if photom!=None]
# Shortcut if no photometry is done
if not kwargs_dict['do_photom']:
photom_attempts = 'Complete'
gc.collect()
return
# Check that all photometry completed
photom_attempts, photom_output_list = CAAPR.CAAPR_Photom.PhotomCheck(photom_attempts, photom_output_list, source_dict, bands_dict, kwargs_dict)
# Record photometry results to file
CAAPR.CAAPR_IO.RecordPhotom(photom_list, source_dict, bands_dict, kwargs_dict)
# Prepare thumbnail images for bands excluded from photometry
CAAPR.CAAPR_Photom.ExcludedThumb(source_dict, bands_dict, kwargs_dict)
# Create grid of thumbnail images
CAAPR.CAAPR_IO.PhotomThumbGrid(source_dict, bands_dict, kwargs_dict)
# Report time taken to do photometry, and tidy up
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Time taken performing actual photometry: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-photom_start,4))+' seconds.'
# Tidy up temporary files and paths
bands_dict = PathTidy(source_dict, bands_dict, kwargs_dict)
# Report time taken for source, and tidy up garbage
gc.collect()
if kwargs_dict['verbose']: print '['+source_dict['name']+'] Total time taken for souce: '+str(ChrisFuncs.FromGitHub.randlet.ToPrecision(time.time()-source_start,4))+' seconds.'
if kwargs_dict['thumbnails']==True and kwargs_dict['messy']==False:
[os.remove(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps',processed_map)) for processed_map in os.listdir(os.path.join(kwargs_dict['temp_dir_path'],'Processed_Maps')) if '.fits' in processed_map]
# Define function to check if data actually exists for any band for this source
def SourcePrelim(source_dict, bands_dict, kwargs_dict):
# Check that any of the bands actually have data for this source
kwargs_dict_copy = copy.deepcopy(kwargs_dict)
kwargs_dict_copy['verbose'] = False
bands_check = []
for band in bands_dict.keys():
source_id = source_dict['name']+'_'+bands_dict[band]['band_name']
in_fitspath, file_found = CAAPR.CAAPR_Pipeline.FilePrelim(source_dict, bands_dict[band], kwargs_dict_copy)
bands_check.append(file_found)
# Report to user if no data found
if True not in bands_check:
|
# Return result
if True not in bands_check:
return False
elif True in bands_check:
return True
# Define function that does basic initial handling of band parameters
def BandInitiate(band_dict):
# Make sure band has content
if band_dict==None:
return band_dict
# Parse band cutout request, converting string to boolean if necessary
if band_dict['make_cutout']=='True':
band_dict['make_cutout']=True
elif band_dict['make_cutout']=='False':
band_dict['make_cutout']=False
else:
try:
band_dict['make_cutout'] = float(band_dict['make_cutout'])
except:
raise Exception('Cutout request not understood; should either be False, or width of cutout in arcseconds.')
# Reset band directory to inviolate value, to purge any holdovers from previous source
band_dict['band_dir'] = band_dict['band_dir_inviolate']
# Return band dict
return band_dict
# Define function that performs preimilary checks of file type and location
def FilePrelim(source_dict, band_dict, kwargs_dict):
# Determine whether the user is specificing a directroy full of FITS files in this band (in which case use standardised filename format), or just a single FITS file
try:
if os.path.isdir(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'], source_dict['name']+'_'+band_dict['band_name'] )
elif os.path.isfile(band_dict['band_dir']):
in_fitspath = os.path.join( band_dict['band_dir'] )
except:
pdb.set_trace()
# Work out whether the file extension for FITS file in question is .fits or .fits.gz
file_found = False
try:
if os.path.exists(in_fitspath+'.fits'):
in_fitspath = in_fitspath+'.fits'
file_found = True
elif os.path.exists(in_fitspath+'.fits.gz'):
in_fitspath = in_fitspath+'.fits.gz'
file_found = True
except:
raise Exception('Path provided for band '+str(band_dict['band_name'])+' refers to neither a file nor a folder.')
# Return file values
return in_fitspath, file_found
# Initiate the pod (Photometry Organisation Dictionary)
def PodInitiate(in_fitspath, source_dict, band_dict, kwargs_dict):
source_id = source_dict['name']+'_'+band_dict['band_name']
if kwargs_dict['verbose']: print '['+source_id+'] Reading in FITS data.'
# Read in FITS file in question
in_fitsdata = astropy.io.fits.open(in_fitspath)
in_image = in_fitsdata[0].data
in_header = in_fitsdata[0].header
in_fitsdata.close()
in_wcs = astropy.wcs.WCS(in_header)
in_fitspath_size = float(os.stat(in_fitspath).st_size)
# Create the pod (Photometry Organisation Dictionary), which will bundle all the photometry data for this source & band into one dictionary to be passed between functions
pod = {'in_fitspath':in_fitspath,
'in_image':in_image,
'in_header':in_header,
'in_wcs':in_wcs,
'cutout':in_image.copy(),
'output_dir_path':kwargs_dict['output_dir_path'],
'temp
|
print '['+source_id+'] No data found in target directory for current source.'
# Make null entries in tables, as necessary
if kwargs_dict['fit_apertures']==True:
null_aperture_combined = [np.NaN, np.NaN, np.NaN, np.NaN]
CAAPR.CAAPR_IO.RecordAperture(null_aperture_combined, source_dict, kwargs_dict)
if kwargs_dict['do_photom']==True:
CAAPR.CAAPR_IO.RecordPhotom([], source_dict, bands_dict, kwargs_dict)
|
conditional_block
|
index.js
|
var btnRule = pageRule.querySelector('.bottom-btn img');
var btnQuestion1 = pageQuestion1.querySelector('.bottom-btn img');
var btnGifPlay = pageQuestion1.querySelector('.btn-play');
var btnGifImg = pageQuestion1.querySelector('.gif-img');
var btnQuestion2 = pageQuestion2.querySelector('.bottom-btn img');
var btnRestart = pageFail.querySelector('.bottom-btn img');
var btnZhizhen = pageReward.querySelector('.zhizhen');
var btnAgain = pageAgain.querySelector('.bottom-btn img');
var btnMessages = pageMessages.querySelector('.bottom-btn img');
var jiangpinImg = pageMessages.querySelector('.jiangpin img');
var jiangpinName = pageMessages.querySelector('.jiangpin p');
var answer1 = '';
var answer2 = {};
var currentMaster = 1;
var reward = '';
//云朵飘动
function yunduo() {
yun1Left++;
yun1.style.left = -yun1Left+'px';
yun2.style.left = (yun2Left-yun1Left)+'px';
if(yun1Left > yun2Left){
yun1Left = 0;
}
requestAnimationFrame(yunduo);
}
//隐藏系统alerturl
window.alert = function(name){
var iframe = document.createElement("IFRAME");
iframe.style.display="none";
iframe.setAttribute("src", 'data:text/plain,');
document.documentElement.appendChild(iframe);
window.frames[0].window.alert(name);
iframe.parentNode.removeChild(iframe);
};
//预加载
var imgArr = [
'img/answer1-gif.gif',
'img/answer1-gif-1.jpg',
'img/img-index.png',
'img/bg-messages.png',
'img/bg-question1.png',
'img/bg-question2.png',
'img/bg-rule.png',
'img/bg-zhuanpan.png',
'img/1.jpg',
'img/2.png',
'img/3.jpg',
'img/4.jpg',
'img/wxicon.jpg'
];
var loading = document.querySelector('.loading');
var loadingPro = loading.querySelector('.top');
var imgArrLength = imgArr.length;
var imageKey = 0;
imgArr.forEach(function (val,key) {
var oImg = new Image();
oImg.onload = function(){
oImg.onload = null;
loadingPro.style.width = Math.ceil(100*(++imageKey)/imgArrLength)+'%';
if (imageKey == imgArrLength) {
$('.preload-bg').each(function (i,v) {
v.style.backgroundImage = 'url('+v.dataset.preload_src+')';
});
$('.preload-img').each(function (i,v) {
v.src = v.dataset.preload_src;
});
loading.classList.add('none');
pageIndex.classList.remove('none');
// pageReward.classList.remove('none');
// pageMessages.classList.remove('none');
requestAnimationFrame(yunduo);
document.addEventListener("WeixinJSBridgeReady", function () {//微信
bgMusic.play();
}, false);
}
};
oImg.src = val;
});
btnStart.onclick = function () {
btnAudio.play();
pageIndex.classList.add('none');
pageRule.classList.remove('none');
};
btnRule.onclick = function () {
btnAudio.play();
pageRule.classList.add('none');
pageQuestion1.classList.remove('none');
};
//问题一选答案并记录
$('#question1 .answer1').each(function (i,val) {
val.onclick =function () {
btnAudio.play();
$('#question1 .answer1').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
answer1 = val.dataset.value;
}
});
//问题一点击按钮播放
btnGifPlay.onclick = function () {
btnAudio.play();
var _this = this;
var src = btnGifImg.src;
var dataSrc = btnGifImg.dataset.src;
_this.classList.add('none');
setTimeout(function () {
btnGifImg.src = dataSrc;
},500);
setTimeout(function () {
_this.classList.remove('none');
btnGifImg.src = src;
},9500);
};
btnQuestion1.onclick = function () {
btnAudio.play();
if (answer1) {
pageQuestion1.classList.add('none');
pageQuestion2.classList.remove('none');
}
};
//问题二选和尚
$('.master-small img').each(function (key,val) {
val.onclick =function () {
btnAudio.play();
$('.bingqi img').each(function (i,v) {
v.classList.remove('active');
});
if (answer2['as'+(key+1)]) {
pageQuestion2.querySelectorAll('.bingqi img')[answer2['as'+(key+1)]-1].classList.add('active');
}
$('.master-small img').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
$('.master-big').each(function (i,v) {
v.classList.add('none');
});
pageQuestion2.querySelectorAll('.master-big')[key].classList.remove('none');
currentMaster = key+1;
}
});
//问题二选兵器
$('.bingqi img').each(function (key,val) {
val.onclick = function () {
btnAudio.play();
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
val.classList.add('active');
answer2['as'+currentMaster] = key+1;
}
});
btnQuestion2.onclick = function () {
btnAudio.play();
if (answer2.as3 && answer2.as1 && answer2.as2) {
if ((answer2.as1 === 1) && (answer2.as2 === 2) && (answer2.as3 === 4) && (answer1=='c')) {
pageQuestion2.classList.add('none');
pageReward.classList.remove('none');
} else {
pageFail.classList.remove('none');
}
}
else {
alert ('注意!三位大师都要匹配相应兵器哦~');
}
};
//重新开始
btnRestart.onclick = function () {
btnAudio.play();
pageFail.classList.add('none');
pageIndex.classList.remove('none');
//清空之前答案
pageQuestion1.querySelectorAll('.answer1').forEach(function (v) {
v.classList.remove('active');
});
answer1 = '';
pageQuestion2.querySelectorAll('.master-small img').forEach(function (v) {
v.classList.remove('active');
});
pageQuestion2.querySelectorAll('.master-small img')[0].classList.add('active');
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
answer2 = {};
currentMaster = 1;
};
//抽奖
var is_click = 1;
btnZhizhen.onclick = function () {
btnAudio.play();
if (!is_click) {
return false;
}
var _this = this;
var zhongjiang = 1;
is_click = 0;
//请求抽奖结果决定停止位置
$.ajax({
type: 'post',
url: 'https://wx.ouu.me/home/getHomeLuckdraw',
data: {
openid: open_id
},
dataType: 'json',
success: function(data){
reward = data.date.yes || '5';
if (reward == '1') {
_this.classList.add('yideng');
jiangpinImg.src = 'img/1.jpg';
jiangpinName.innerHTML = '昂坪360 精美行李带<span>(价值HK$120)</span>';
} else if(reward == '2') {
var erdeng = (Math.random()>0.5)? 'erdeng1':'erdeng2';
_this.classList.add(erdeng);
jiangpinImg.src = 'img/2.png';
jiangpinName.innerHTML = '昂坪360 化妆品收纳袋<span>(价值HK$120)</span>';
} else if(reward == '3') {
var sandeng = (Math.random()>0.5)? 'sandeng1':'sandeng2';
_this.classList.add(sandeng);
jiangpinImg.src = 'img/3.jpg';
jiangpinName.innerHTML = '昂坪360 缆车小钱包<span>(价值HK$130)</span>';
} else if(reward == '4') {
var youxiu = (Math.random()>0.5)? 'youxiu1':'youxiu2';
_this.classList.add(youxiu);
jiangpinImg.src = 'img/4.jpg';
jiangpinName.innerHTML = '驴妈妈 小驴公仔';
} else {
var xiexie = (Math.random()>0.5)? 'xiexie1':'xiexie2';
_this.classList.add(xiexie);
zhongjiang = 0
}
setTimeout
|
var yun1Left = 0;
var yun2Left = parseInt(window.getComputedStyle(yun2).left);
var btnStart = pageIndex.querySelector('.bottom-btn img');
|
random_line_split
|
|
index.js
|
ifImg = pageQuestion1.querySelector('.gif-img');
var btnQuestion2 = pageQuestion2.querySelector('.bottom-btn img');
var btnRestart = pageFail.querySelector('.bottom-btn img');
var btnZhizhen = pageReward.querySelector('.zhizhen');
var btnAgain = pageAgain.querySelector('.bottom-btn img');
var btnMessages = pageMessages.querySelector('.bottom-btn img');
var jiangpinImg = pageMessages.querySelector('.jiangpin img');
var jiangpinName = pageMessages.querySelector('.jiangpin p');
var answer1 = '';
var answer2 = {};
var currentMaster = 1;
var reward = '';
//云朵飘动
function yunduo()
|
yun1Left++;
yun1.style.left = -yun1Left+'px';
yun2.style.left = (yun2Left-yun1Left)+'px';
if(yun1Left > yun2Left){
yun1Left = 0;
}
requestAnimationFrame(yunduo);
}
//隐藏系统alerturl
window.alert = function(name){
var iframe = document.createElement("IFRAME");
iframe.style.display="none";
iframe.setAttribute("src", 'data:text/plain,');
document.documentElement.appendChild(iframe);
window.frames[0].window.alert(name);
iframe.parentNode.removeChild(iframe);
};
//预加载
var imgArr = [
'img/answer1-gif.gif',
'img/answer1-gif-1.jpg',
'img/img-index.png',
'img/bg-messages.png',
'img/bg-question1.png',
'img/bg-question2.png',
'img/bg-rule.png',
'img/bg-zhuanpan.png',
'img/1.jpg',
'img/2.png',
'img/3.jpg',
'img/4.jpg',
'img/wxicon.jpg'
];
var loading = document.querySelector('.loading');
var loadingPro = loading.querySelector('.top');
var imgArrLength = imgArr.length;
var imageKey = 0;
imgArr.forEach(function (val,key) {
var oImg = new Image();
oImg.onload = function(){
oImg.onload = null;
loadingPro.style.width = Math.ceil(100*(++imageKey)/imgArrLength)+'%';
if (imageKey == imgArrLength) {
$('.preload-bg').each(function (i,v) {
v.style.backgroundImage = 'url('+v.dataset.preload_src+')';
});
$('.preload-img').each(function (i,v) {
v.src = v.dataset.preload_src;
});
loading.classList.add('none');
pageIndex.classList.remove('none');
// pageReward.classList.remove('none');
// pageMessages.classList.remove('none');
requestAnimationFrame(yunduo);
document.addEventListener("WeixinJSBridgeReady", function () {//微信
bgMusic.play();
}, false);
}
};
oImg.src = val;
});
btnStart.onclick = function () {
btnAudio.play();
pageIndex.classList.add('none');
pageRule.classList.remove('none');
};
btnRule.onclick = function () {
btnAudio.play();
pageRule.classList.add('none');
pageQuestion1.classList.remove('none');
};
//问题一选答案并记录
$('#question1 .answer1').each(function (i,val) {
val.onclick =function () {
btnAudio.play();
$('#question1 .answer1').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
answer1 = val.dataset.value;
}
});
//问题一点击按钮播放
btnGifPlay.onclick = function () {
btnAudio.play();
var _this = this;
var src = btnGifImg.src;
var dataSrc = btnGifImg.dataset.src;
_this.classList.add('none');
setTimeout(function () {
btnGifImg.src = dataSrc;
},500);
setTimeout(function () {
_this.classList.remove('none');
btnGifImg.src = src;
},9500);
};
btnQuestion1.onclick = function () {
btnAudio.play();
if (answer1) {
pageQuestion1.classList.add('none');
pageQuestion2.classList.remove('none');
}
};
//问题二选和尚
$('.master-small img').each(function (key,val) {
val.onclick =function () {
btnAudio.play();
$('.bingqi img').each(function (i,v) {
v.classList.remove('active');
});
if (answer2['as'+(key+1)]) {
pageQuestion2.querySelectorAll('.bingqi img')[answer2['as'+(key+1)]-1].classList.add('active');
}
$('.master-small img').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
$('.master-big').each(function (i,v) {
v.classList.add('none');
});
pageQuestion2.querySelectorAll('.master-big')[key].classList.remove('none');
currentMaster = key+1;
}
});
//问题二选兵器
$('.bingqi img').each(function (key,val) {
val.onclick = function () {
btnAudio.play();
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
val.classList.add('active');
answer2['as'+currentMaster] = key+1;
}
});
btnQuestion2.onclick = function () {
btnAudio.play();
if (answer2.as3 && answer2.as1 && answer2.as2) {
if ((answer2.as1 === 1) && (answer2.as2 === 2) && (answer2.as3 === 4) && (answer1=='c')) {
pageQuestion2.classList.add('none');
pageReward.classList.remove('none');
} else {
pageFail.classList.remove('none');
}
}
else {
alert ('注意!三位大师都要匹配相应兵器哦~');
}
};
//重新开始
btnRestart.onclick = function () {
btnAudio.play();
pageFail.classList.add('none');
pageIndex.classList.remove('none');
//清空之前答案
pageQuestion1.querySelectorAll('.answer1').forEach(function (v) {
v.classList.remove('active');
});
answer1 = '';
pageQuestion2.querySelectorAll('.master-small img').forEach(function (v) {
v.classList.remove('active');
});
pageQuestion2.querySelectorAll('.master-small img')[0].classList.add('active');
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
answer2 = {};
currentMaster = 1;
};
//抽奖
var is_click = 1;
btnZhizhen.onclick = function () {
btnAudio.play();
if (!is_click) {
return false;
}
var _this = this;
var zhongjiang = 1;
is_click = 0;
//请求抽奖结果决定停止位置
$.ajax({
type: 'post',
url: 'https://wx.ouu.me/home/getHomeLuckdraw',
data: {
openid: open_id
},
dataType: 'json',
success: function(data){
reward = data.date.yes || '5';
if (reward == '1') {
_this.classList.add('yideng');
jiangpinImg.src = 'img/1.jpg';
jiangpinName.innerHTML = '昂坪360 精美行李带<span>(价值HK$120)</span>';
} else if(reward == '2') {
var erdeng = (Math.random()>0.5)? 'erdeng1':'erdeng2';
_this.classList.add(erdeng);
jiangpinImg.src = 'img/2.png';
jiangpinName.innerHTML = '昂坪360 化妆品收纳袋<span>(价值HK$120)</span>';
} else if(reward == '3') {
var sandeng = (Math.random()>0.5)? 'sandeng1':'sandeng2';
_this.classList.add(sandeng);
jiangpinImg.src = 'img/3.jpg';
jiangpinName.innerHTML = '昂坪360 缆车小钱包<span>(价值HK$130)</span>';
} else if(reward == '4') {
var youxiu = (Math.random()>0.5)? 'youxiu1':'youxiu2';
_this.classList.add(youxiu);
jiangpinImg.src = 'img/4.jpg';
jiangpinName.innerHTML = '驴妈妈 小驴公仔';
} else {
var xiexie = (Math.random()>0.5)? 'xiexie1':'xiexie2';
_this.classList.add(xiexie);
zhongjiang = 0
}
setTimeout(function () {
if (zhongjiang) {
pageReward.classList.add('none');
pageMessages.classList.remove('none');
} else {
pageAgain.classList.remove('none');
}
},5200)
}
});
};
//再来一次
btnAgain.onclick = function () {
btnAudio.play();
btnZhizhen.className = 'zhizhen';
is_click
|
{
|
identifier_name
|
index.js
|
ifImg = pageQuestion1.querySelector('.gif-img');
var btnQuestion2 = pageQuestion2.querySelector('.bottom-btn img');
var btnRestart = pageFail.querySelector('.bottom-btn img');
var btnZhizhen = pageReward.querySelector('.zhizhen');
var btnAgain = pageAgain.querySelector('.bottom-btn img');
var btnMessages = pageMessages.querySelector('.bottom-btn img');
var jiangpinImg = pageMessages.querySelector('.jiangpin img');
var jiangpinName = pageMessages.querySelector('.jiangpin p');
var answer1 = '';
var answer2 = {};
var currentMaster = 1;
var reward = '';
//云朵飘动
function yunduo() {
yu
|
系统alerturl
window.alert = function(name){
var iframe = document.createElement("IFRAME");
iframe.style.display="none";
iframe.setAttribute("src", 'data:text/plain,');
document.documentElement.appendChild(iframe);
window.frames[0].window.alert(name);
iframe.parentNode.removeChild(iframe);
};
//预加载
var imgArr = [
'img/answer1-gif.gif',
'img/answer1-gif-1.jpg',
'img/img-index.png',
'img/bg-messages.png',
'img/bg-question1.png',
'img/bg-question2.png',
'img/bg-rule.png',
'img/bg-zhuanpan.png',
'img/1.jpg',
'img/2.png',
'img/3.jpg',
'img/4.jpg',
'img/wxicon.jpg'
];
var loading = document.querySelector('.loading');
var loadingPro = loading.querySelector('.top');
var imgArrLength = imgArr.length;
var imageKey = 0;
imgArr.forEach(function (val,key) {
var oImg = new Image();
oImg.onload = function(){
oImg.onload = null;
loadingPro.style.width = Math.ceil(100*(++imageKey)/imgArrLength)+'%';
if (imageKey == imgArrLength) {
$('.preload-bg').each(function (i,v) {
v.style.backgroundImage = 'url('+v.dataset.preload_src+')';
});
$('.preload-img').each(function (i,v) {
v.src = v.dataset.preload_src;
});
loading.classList.add('none');
pageIndex.classList.remove('none');
// pageReward.classList.remove('none');
// pageMessages.classList.remove('none');
requestAnimationFrame(yunduo);
document.addEventListener("WeixinJSBridgeReady", function () {//微信
bgMusic.play();
}, false);
}
};
oImg.src = val;
});
btnStart.onclick = function () {
btnAudio.play();
pageIndex.classList.add('none');
pageRule.classList.remove('none');
};
btnRule.onclick = function () {
btnAudio.play();
pageRule.classList.add('none');
pageQuestion1.classList.remove('none');
};
//问题一选答案并记录
$('#question1 .answer1').each(function (i,val) {
val.onclick =function () {
btnAudio.play();
$('#question1 .answer1').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
answer1 = val.dataset.value;
}
});
//问题一点击按钮播放
btnGifPlay.onclick = function () {
btnAudio.play();
var _this = this;
var src = btnGifImg.src;
var dataSrc = btnGifImg.dataset.src;
_this.classList.add('none');
setTimeout(function () {
btnGifImg.src = dataSrc;
},500);
setTimeout(function () {
_this.classList.remove('none');
btnGifImg.src = src;
},9500);
};
btnQuestion1.onclick = function () {
btnAudio.play();
if (answer1) {
pageQuestion1.classList.add('none');
pageQuestion2.classList.remove('none');
}
};
//问题二选和尚
$('.master-small img').each(function (key,val) {
val.onclick =function () {
btnAudio.play();
$('.bingqi img').each(function (i,v) {
v.classList.remove('active');
});
if (answer2['as'+(key+1)]) {
pageQuestion2.querySelectorAll('.bingqi img')[answer2['as'+(key+1)]-1].classList.add('active');
}
$('.master-small img').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
$('.master-big').each(function (i,v) {
v.classList.add('none');
});
pageQuestion2.querySelectorAll('.master-big')[key].classList.remove('none');
currentMaster = key+1;
}
});
//问题二选兵器
$('.bingqi img').each(function (key,val) {
val.onclick = function () {
btnAudio.play();
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
val.classList.add('active');
answer2['as'+currentMaster] = key+1;
}
});
btnQuestion2.onclick = function () {
btnAudio.play();
if (answer2.as3 && answer2.as1 && answer2.as2) {
if ((answer2.as1 === 1) && (answer2.as2 === 2) && (answer2.as3 === 4) && (answer1=='c')) {
pageQuestion2.classList.add('none');
pageReward.classList.remove('none');
} else {
pageFail.classList.remove('none');
}
}
else {
alert ('注意!三位大师都要匹配相应兵器哦~');
}
};
//重新开始
btnRestart.onclick = function () {
btnAudio.play();
pageFail.classList.add('none');
pageIndex.classList.remove('none');
//清空之前答案
pageQuestion1.querySelectorAll('.answer1').forEach(function (v) {
v.classList.remove('active');
});
answer1 = '';
pageQuestion2.querySelectorAll('.master-small img').forEach(function (v) {
v.classList.remove('active');
});
pageQuestion2.querySelectorAll('.master-small img')[0].classList.add('active');
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
answer2 = {};
currentMaster = 1;
};
//抽奖
var is_click = 1;
btnZhizhen.onclick = function () {
btnAudio.play();
if (!is_click) {
return false;
}
var _this = this;
var zhongjiang = 1;
is_click = 0;
//请求抽奖结果决定停止位置
$.ajax({
type: 'post',
url: 'https://wx.ouu.me/home/getHomeLuckdraw',
data: {
openid: open_id
},
dataType: 'json',
success: function(data){
reward = data.date.yes || '5';
if (reward == '1') {
_this.classList.add('yideng');
jiangpinImg.src = 'img/1.jpg';
jiangpinName.innerHTML = '昂坪360 精美行李带<span>(价值HK$120)</span>';
} else if(reward == '2') {
var erdeng = (Math.random()>0.5)? 'erdeng1':'erdeng2';
_this.classList.add(erdeng);
jiangpinImg.src = 'img/2.png';
jiangpinName.innerHTML = '昂坪360 化妆品收纳袋<span>(价值HK$120)</span>';
} else if(reward == '3') {
var sandeng = (Math.random()>0.5)? 'sandeng1':'sandeng2';
_this.classList.add(sandeng);
jiangpinImg.src = 'img/3.jpg';
jiangpinName.innerHTML = '昂坪360 缆车小钱包<span>(价值HK$130)</span>';
} else if(reward == '4') {
var youxiu = (Math.random()>0.5)? 'youxiu1':'youxiu2';
_this.classList.add(youxiu);
jiangpinImg.src = 'img/4.jpg';
jiangpinName.innerHTML = '驴妈妈 小驴公仔';
} else {
var xiexie = (Math.random()>0.5)? 'xiexie1':'xiexie2';
_this.classList.add(xiexie);
zhongjiang = 0
}
setTimeout(function () {
if (zhongjiang) {
pageReward.classList.add('none');
pageMessages.classList.remove('none');
} else {
pageAgain.classList.remove('none');
}
},5200)
}
});
};
//再来一次
btnAgain.onclick = function () {
btnAudio.play();
btnZhizhen.className = 'zhizhen';
is
|
n1Left++;
yun1.style.left = -yun1Left+'px';
yun2.style.left = (yun2Left-yun1Left)+'px';
if(yun1Left > yun2Left){
yun1Left = 0;
}
requestAnimationFrame(yunduo);
}
//隐藏
|
identifier_body
|
index.js
|
ifImg = pageQuestion1.querySelector('.gif-img');
var btnQuestion2 = pageQuestion2.querySelector('.bottom-btn img');
var btnRestart = pageFail.querySelector('.bottom-btn img');
var btnZhizhen = pageReward.querySelector('.zhizhen');
var btnAgain = pageAgain.querySelector('.bottom-btn img');
var btnMessages = pageMessages.querySelector('.bottom-btn img');
var jiangpinImg = pageMessages.querySelector('.jiangpin img');
var jiangpinName = pageMessages.querySelector('.jiangpin p');
var answer1 = '';
var answer2 = {};
var currentMaster = 1;
var reward = '';
//云朵飘动
function yunduo() {
yun1Left++;
yun1.style.left = -yun1Left+'px';
yun2.style.left = (yun2Left-yun1Left)+'px';
if(yun1Left > yun2Left){
|
uestAnimationFrame(yunduo);
}
//隐藏系统alerturl
window.alert = function(name){
var iframe = document.createElement("IFRAME");
iframe.style.display="none";
iframe.setAttribute("src", 'data:text/plain,');
document.documentElement.appendChild(iframe);
window.frames[0].window.alert(name);
iframe.parentNode.removeChild(iframe);
};
//预加载
var imgArr = [
'img/answer1-gif.gif',
'img/answer1-gif-1.jpg',
'img/img-index.png',
'img/bg-messages.png',
'img/bg-question1.png',
'img/bg-question2.png',
'img/bg-rule.png',
'img/bg-zhuanpan.png',
'img/1.jpg',
'img/2.png',
'img/3.jpg',
'img/4.jpg',
'img/wxicon.jpg'
];
var loading = document.querySelector('.loading');
var loadingPro = loading.querySelector('.top');
var imgArrLength = imgArr.length;
var imageKey = 0;
imgArr.forEach(function (val,key) {
var oImg = new Image();
oImg.onload = function(){
oImg.onload = null;
loadingPro.style.width = Math.ceil(100*(++imageKey)/imgArrLength)+'%';
if (imageKey == imgArrLength) {
$('.preload-bg').each(function (i,v) {
v.style.backgroundImage = 'url('+v.dataset.preload_src+')';
});
$('.preload-img').each(function (i,v) {
v.src = v.dataset.preload_src;
});
loading.classList.add('none');
pageIndex.classList.remove('none');
// pageReward.classList.remove('none');
// pageMessages.classList.remove('none');
requestAnimationFrame(yunduo);
document.addEventListener("WeixinJSBridgeReady", function () {//微信
bgMusic.play();
}, false);
}
};
oImg.src = val;
});
btnStart.onclick = function () {
btnAudio.play();
pageIndex.classList.add('none');
pageRule.classList.remove('none');
};
btnRule.onclick = function () {
btnAudio.play();
pageRule.classList.add('none');
pageQuestion1.classList.remove('none');
};
//问题一选答案并记录
$('#question1 .answer1').each(function (i,val) {
val.onclick =function () {
btnAudio.play();
$('#question1 .answer1').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
answer1 = val.dataset.value;
}
});
//问题一点击按钮播放
btnGifPlay.onclick = function () {
btnAudio.play();
var _this = this;
var src = btnGifImg.src;
var dataSrc = btnGifImg.dataset.src;
_this.classList.add('none');
setTimeout(function () {
btnGifImg.src = dataSrc;
},500);
setTimeout(function () {
_this.classList.remove('none');
btnGifImg.src = src;
},9500);
};
btnQuestion1.onclick = function () {
btnAudio.play();
if (answer1) {
pageQuestion1.classList.add('none');
pageQuestion2.classList.remove('none');
}
};
//问题二选和尚
$('.master-small img').each(function (key,val) {
val.onclick =function () {
btnAudio.play();
$('.bingqi img').each(function (i,v) {
v.classList.remove('active');
});
if (answer2['as'+(key+1)]) {
pageQuestion2.querySelectorAll('.bingqi img')[answer2['as'+(key+1)]-1].classList.add('active');
}
$('.master-small img').each(function (i,v) {
v.classList.remove('active');
});
val.classList.add('active');
$('.master-big').each(function (i,v) {
v.classList.add('none');
});
pageQuestion2.querySelectorAll('.master-big')[key].classList.remove('none');
currentMaster = key+1;
}
});
//问题二选兵器
$('.bingqi img').each(function (key,val) {
val.onclick = function () {
btnAudio.play();
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
val.classList.add('active');
answer2['as'+currentMaster] = key+1;
}
});
btnQuestion2.onclick = function () {
btnAudio.play();
if (answer2.as3 && answer2.as1 && answer2.as2) {
if ((answer2.as1 === 1) && (answer2.as2 === 2) && (answer2.as3 === 4) && (answer1=='c')) {
pageQuestion2.classList.add('none');
pageReward.classList.remove('none');
} else {
pageFail.classList.remove('none');
}
}
else {
alert ('注意!三位大师都要匹配相应兵器哦~');
}
};
//重新开始
btnRestart.onclick = function () {
btnAudio.play();
pageFail.classList.add('none');
pageIndex.classList.remove('none');
//清空之前答案
pageQuestion1.querySelectorAll('.answer1').forEach(function (v) {
v.classList.remove('active');
});
answer1 = '';
pageQuestion2.querySelectorAll('.master-small img').forEach(function (v) {
v.classList.remove('active');
});
pageQuestion2.querySelectorAll('.master-small img')[0].classList.add('active');
pageQuestion2.querySelectorAll('.bingqi img').forEach(function (v) {
v.classList.remove('active');
});
answer2 = {};
currentMaster = 1;
};
//抽奖
var is_click = 1;
btnZhizhen.onclick = function () {
btnAudio.play();
if (!is_click) {
return false;
}
var _this = this;
var zhongjiang = 1;
is_click = 0;
//请求抽奖结果决定停止位置
$.ajax({
type: 'post',
url: 'https://wx.ouu.me/home/getHomeLuckdraw',
data: {
openid: open_id
},
dataType: 'json',
success: function(data){
reward = data.date.yes || '5';
if (reward == '1') {
_this.classList.add('yideng');
jiangpinImg.src = 'img/1.jpg';
jiangpinName.innerHTML = '昂坪360 精美行李带<span>(价值HK$120)</span>';
} else if(reward == '2') {
var erdeng = (Math.random()>0.5)? 'erdeng1':'erdeng2';
_this.classList.add(erdeng);
jiangpinImg.src = 'img/2.png';
jiangpinName.innerHTML = '昂坪360 化妆品收纳袋<span>(价值HK$120)</span>';
} else if(reward == '3') {
var sandeng = (Math.random()>0.5)? 'sandeng1':'sandeng2';
_this.classList.add(sandeng);
jiangpinImg.src = 'img/3.jpg';
jiangpinName.innerHTML = '昂坪360 缆车小钱包<span>(价值HK$130)</span>';
} else if(reward == '4') {
var youxiu = (Math.random()>0.5)? 'youxiu1':'youxiu2';
_this.classList.add(youxiu);
jiangpinImg.src = 'img/4.jpg';
jiangpinName.innerHTML = '驴妈妈 小驴公仔';
} else {
var xiexie = (Math.random()>0.5)? 'xiexie1':'xiexie2';
_this.classList.add(xiexie);
zhongjiang = 0
}
setTimeout(function () {
if (zhongjiang) {
pageReward.classList.add('none');
pageMessages.classList.remove('none');
} else {
pageAgain.classList.remove('none');
}
},5200)
}
});
};
//再来一次
btnAgain.onclick = function () {
btnAudio.play();
btnZhizhen.className = 'zhizhen';
is_click
|
yun1Left = 0;
}
req
|
conditional_block
|
images.py
|
)
# timestamp of when image was added
# images:ids:timestamps = sorted (ids,timestamp)
# all the image ids for the page
# images:page_ids:<page_url> (ids)
# last time an image was added from page
# images:pages:timestamps = sorted (url,timestamp)
# images meta data
# images:id = {}
def _image_to_dict(self, image):
data = {}
ignored_attrs = ['data']
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
if attr in ignored_attrs:
continue
v = getattr(image,attr)
if v is not None:
data[attr] = v
return data
def _dict_to_image(self, data):
image = o.Image()
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
v = data.get(attr)
if v is not None:
# we might need to update the value
# type, since all values come back
# from redis as strings
attr_type = attrs[1]
# float
if attr_type == 4:
setattr(image,attr,float(v))
# int
elif attr_type == 8:
setattr(image,attr,int(v))
else:
setattr(image,attr,v)
return image
def _delete_from_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# remove it from the id set
pipe.zrem('images:ids:timestamps',image.id)
# remove it's hash
pipe.delete('images:%s' % image.id)
# decriment the count for it's image data
pipe.srem('images:datainstances:%s' % image.shahash,
image.id)
# remove image from the page's id set
if image.source_page_url:
pipe.zrem('images:page_ids:%s' % image.source_page_url,
image.id)
# make it happen
pipe.execute()
return True
def _save_to_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# if our image doesn't have an id, set it up w/ one
if not image.id:
print 'got new image: %s' % image.shahash
image.id = self.rc.incr('images:next_id')
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# check and see if we used to have a different shahash
old_shahash = self.rc.hget('images:%s' % image.id,'shahash')
if old_shahash != image.shahash:
# remove our id from the old shahash tracker
pipe.srem('images:datainstances:%s' % old_shahash,
image.id)
# add it to the new tracker
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# update / set our timestamp
da = 0.0
if image.downloaded_at:
da = image.downloaded_at
else:
da = time.time()
pipe.zadd('images:ids:timestamps',image.id, da)
# add this image to the page's id set
if image.source_page_url:
pipe.zadd('images:page_ids:%s' % image.source_page_url,
image.id, da)
# update our last scrape time for the page
pipe.zadd('images:pages:timestamps',
image.source_page_url, image.id)
# take our image and make a dict
image_data = self._image_to_dict(image)
# set our data to redis
key = 'images:%s' % image.id
pipe.hmset(key,image_data)
# execute our pipe
pipe.execute()
return image
def _get_from_redis(self, image_id):
# if the image id is in the id set than pull it's details
if self.rc.zrank('images:ids:timestamps',image_id) is not None:
# get the image data from redis
key = 'images:%s' % image_id
image_data = self.rc.hgetall(key)
if not image_data:
print 'redis had no image data'
return None
image = self._dict_to_image(image_data)
return image
return None
def _populate_image_data(self, image):
if not image.shahash:
return None
with connect(Blobby) as c:
image.data = c.get_data(image.shahash)
return image
def _set_image_data(self, image):
if image.data is not None:
|
return image
def get_image(self, image_id):
""" returns Image for given id or blank Image """
# see if we have an image
image = self._get_from_redis(image_id)
if not image:
raise o.ImageNotFound('Could not get image', image_id)
# pull the actual image data
self._populate_image_data(image)
return image
def add_image(self, image):
""" like set but if we already have this image from this
page we're not going to add it again. will also
fill out image stats (size, dimension) """
# we're only for new images, no i'ds allowed
# if u want to set an id by hand use set_image
if image.id:
raise o.Exception('Can not add image with id')
if not image.data:
raise o.Exception('Image must have data')
if not image.source_page_url:
raise o.Exception('Image must have source page url')
# update it's stats
image = self.populate_image_stats(image)
# only add the image if we haven't seen it beforeQ
# if we've seen it before there will be an id which
# the set of images w/ this data and from this page share
ids = self.rc.sinter('images:datainstance:%s' % image.shahash,
'images:page_ids:%s' % image.source_page_url)
# we don't need to continue
# we'll return back their original msg, w/o the id set
if ids:
print 'image already exists [%s], not setting' % ids
return image
# so the image appears to be new, good for it
return self.set_image(image)
def set_image(self, image):
""" sets image data, returns image """
# would be better if we only saved if it didn't exist
if image.data:
# save the images data
self._set_image_data(image)
# could be an update, could be new
image = self._save_to_redis(image)
# let the world know we have added a new image
self.revent.fire('image_added',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
return image
def delete_image(self, image_id):
""" removes an image """
# get it's image obj
try:
image = self.get_image(image_id)
except o.ImageNotFound, ex:
return False
# delete the redis data
self._delete_from_redis(image)
# see if we need to remove the image data
if self.rc.scard('images:datainstances:%s' % image.shahash) == 0:
# no more images w/ the same data, remove image data
with connect(Blobby) as c:
c.delete_data(image.shahash)
# it's gone, let'm know
self.revent.fire('image_deleted',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
# and we're done!
return True
def get_images_since(self, image_id=None, timestamp=None,
offset=10, limit=0):
""" returns list of tublr images or blank list which were
added after given image id or timestamp """
print '%s %s %s %s' % (image_id,timestamp,limit,offset)
if image_id is not None:
print 'got image id'
# figure out what the current id is and than grab
# our sorted set by index assuming that all ids
# contain an image
next_id = int(self.rc.get('images:next_id') or 0)
# how far from the end is the id given
d = next_id - image_id
start = next_id - d
end = next_id - d + limit - 1
print 'getting between %s %s' % (start,end)
# starting back where we think this image is to + limit
ids = self.rc.zrange('images:ids:timestamps',start,end)
print 'got ids: %
|
with connect(Blobby) as c:
image.shahash = c.set_data(image.data)
|
conditional_block
|
images.py
|
ids:timestamps',image.id)
# remove it's hash
pipe.delete('images:%s' % image.id)
# decriment the count for it's image data
pipe.srem('images:datainstances:%s' % image.shahash,
image.id)
# remove image from the page's id set
if image.source_page_url:
pipe.zrem('images:page_ids:%s' % image.source_page_url,
image.id)
# make it happen
pipe.execute()
return True
def _save_to_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# if our image doesn't have an id, set it up w/ one
if not image.id:
print 'got new image: %s' % image.shahash
image.id = self.rc.incr('images:next_id')
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# check and see if we used to have a different shahash
old_shahash = self.rc.hget('images:%s' % image.id,'shahash')
if old_shahash != image.shahash:
# remove our id from the old shahash tracker
pipe.srem('images:datainstances:%s' % old_shahash,
image.id)
# add it to the new tracker
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# update / set our timestamp
da = 0.0
if image.downloaded_at:
da = image.downloaded_at
else:
da = time.time()
pipe.zadd('images:ids:timestamps',image.id, da)
# add this image to the page's id set
if image.source_page_url:
pipe.zadd('images:page_ids:%s' % image.source_page_url,
image.id, da)
# update our last scrape time for the page
pipe.zadd('images:pages:timestamps',
image.source_page_url, image.id)
# take our image and make a dict
image_data = self._image_to_dict(image)
# set our data to redis
key = 'images:%s' % image.id
pipe.hmset(key,image_data)
# execute our pipe
pipe.execute()
return image
def _get_from_redis(self, image_id):
# if the image id is in the id set than pull it's details
if self.rc.zrank('images:ids:timestamps',image_id) is not None:
# get the image data from redis
key = 'images:%s' % image_id
image_data = self.rc.hgetall(key)
if not image_data:
print 'redis had no image data'
return None
image = self._dict_to_image(image_data)
return image
return None
def _populate_image_data(self, image):
if not image.shahash:
return None
with connect(Blobby) as c:
image.data = c.get_data(image.shahash)
return image
def _set_image_data(self, image):
if image.data is not None:
with connect(Blobby) as c:
image.shahash = c.set_data(image.data)
return image
def get_image(self, image_id):
""" returns Image for given id or blank Image """
# see if we have an image
image = self._get_from_redis(image_id)
if not image:
raise o.ImageNotFound('Could not get image', image_id)
# pull the actual image data
self._populate_image_data(image)
return image
def add_image(self, image):
""" like set but if we already have this image from this
page we're not going to add it again. will also
fill out image stats (size, dimension) """
# we're only for new images, no i'ds allowed
# if u want to set an id by hand use set_image
if image.id:
raise o.Exception('Can not add image with id')
if not image.data:
raise o.Exception('Image must have data')
if not image.source_page_url:
raise o.Exception('Image must have source page url')
# update it's stats
image = self.populate_image_stats(image)
# only add the image if we haven't seen it beforeQ
# if we've seen it before there will be an id which
# the set of images w/ this data and from this page share
ids = self.rc.sinter('images:datainstance:%s' % image.shahash,
'images:page_ids:%s' % image.source_page_url)
# we don't need to continue
# we'll return back their original msg, w/o the id set
if ids:
print 'image already exists [%s], not setting' % ids
return image
# so the image appears to be new, good for it
return self.set_image(image)
def set_image(self, image):
""" sets image data, returns image """
# would be better if we only saved if it didn't exist
if image.data:
# save the images data
self._set_image_data(image)
# could be an update, could be new
image = self._save_to_redis(image)
# let the world know we have added a new image
self.revent.fire('image_added',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
return image
def delete_image(self, image_id):
""" removes an image """
# get it's image obj
try:
image = self.get_image(image_id)
except o.ImageNotFound, ex:
return False
# delete the redis data
self._delete_from_redis(image)
# see if we need to remove the image data
if self.rc.scard('images:datainstances:%s' % image.shahash) == 0:
# no more images w/ the same data, remove image data
with connect(Blobby) as c:
c.delete_data(image.shahash)
# it's gone, let'm know
self.revent.fire('image_deleted',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
# and we're done!
return True
def get_images_since(self, image_id=None, timestamp=None,
offset=10, limit=0):
""" returns list of tublr images or blank list which were
added after given image id or timestamp """
print '%s %s %s %s' % (image_id,timestamp,limit,offset)
if image_id is not None:
print 'got image id'
# figure out what the current id is and than grab
# our sorted set by index assuming that all ids
# contain an image
next_id = int(self.rc.get('images:next_id') or 0)
# how far from the end is the id given
d = next_id - image_id
start = next_id - d
end = next_id - d + limit - 1
print 'getting between %s %s' % (start,end)
# starting back where we think this image is to + limit
ids = self.rc.zrange('images:ids:timestamps',start,end)
print 'got ids: %s' % ids
elif timestamp:
print 'from timestamp: %s' % timestamp
# get ids from our sorted set by it's weight (aka timestamp)
# TODO: not use inf
ids = self.rc.zrangebyscore('images:ids:timestamps',
timestamp,'+inf')
else:
print 'could not find images'
ids = []
# page ids
if offset < len(ids):
ids = ids[offset:max(len(ids),limit)]
else:
ids = []
print 'found ids: %s' % ids
# return images for each ID
images = map(self._get_from_redis,ids)
# populate image data
map(self._populate_image_data,images)
return images
def search(self, source_blog_url=None, since_timestamp=None,
before_timestamp=None, ids=[], source_url=None):
""" returns list of images, searches passed on passed params """
pass
def populate_image_stats(self, image):
""" returns a Image w/ image data + stats filled
out """
ti = image
image_data = ti.data
if not ti.data:
return ti
ti.size = len(image_data)
try:
with connect(Blobby) as c:
ti.shahash = c.get_data_bhash(image_data)
except o.Exception, ex:
|
raise o.Exception('oException getting shahash: %s' % ex.msg)
except Exception, ex:
raise o.Exception('Exception getting shahash: %s' % ex)
|
random_line_split
|
|
images.py
|
)
# timestamp of when image was added
# images:ids:timestamps = sorted (ids,timestamp)
# all the image ids for the page
# images:page_ids:<page_url> (ids)
# last time an image was added from page
# images:pages:timestamps = sorted (url,timestamp)
# images meta data
# images:id = {}
def _image_to_dict(self, image):
data = {}
ignored_attrs = ['data']
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
if attr in ignored_attrs:
continue
v = getattr(image,attr)
if v is not None:
data[attr] = v
return data
def _dict_to_image(self, data):
image = o.Image()
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
v = data.get(attr)
if v is not None:
# we might need to update the value
# type, since all values come back
# from redis as strings
attr_type = attrs[1]
# float
if attr_type == 4:
setattr(image,attr,float(v))
# int
elif attr_type == 8:
setattr(image,attr,int(v))
else:
setattr(image,attr,v)
return image
def _delete_from_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# remove it from the id set
pipe.zrem('images:ids:timestamps',image.id)
# remove it's hash
pipe.delete('images:%s' % image.id)
# decriment the count for it's image data
pipe.srem('images:datainstances:%s' % image.shahash,
image.id)
# remove image from the page's id set
if image.source_page_url:
pipe.zrem('images:page_ids:%s' % image.source_page_url,
image.id)
# make it happen
pipe.execute()
return True
def _save_to_redis(self, image):
# make these a transaction
|
# update / set our timestamp
da = 0.0
if image.downloaded_at:
da = image.downloaded_at
else:
da = time.time()
pipe.zadd('images:ids:timestamps',image.id, da)
# add this image to the page's id set
if image.source_page_url:
pipe.zadd('images:page_ids:%s' % image.source_page_url,
image.id, da)
# update our last scrape time for the page
pipe.zadd('images:pages:timestamps',
image.source_page_url, image.id)
# take our image and make a dict
image_data = self._image_to_dict(image)
# set our data to redis
key = 'images:%s' % image.id
pipe.hmset(key,image_data)
# execute our pipe
pipe.execute()
return image
def _get_from_redis(self, image_id):
# if the image id is in the id set than pull it's details
if self.rc.zrank('images:ids:timestamps',image_id) is not None:
# get the image data from redis
key = 'images:%s' % image_id
image_data = self.rc.hgetall(key)
if not image_data:
print 'redis had no image data'
return None
image = self._dict_to_image(image_data)
return image
return None
def _populate_image_data(self, image):
if not image.shahash:
return None
with connect(Blobby) as c:
image.data = c.get_data(image.shahash)
return image
def _set_image_data(self, image):
if image.data is not None:
with connect(Blobby) as c:
image.shahash = c.set_data(image.data)
return image
def get_image(self, image_id):
""" returns Image for given id or blank Image """
# see if we have an image
image = self._get_from_redis(image_id)
if not image:
raise o.ImageNotFound('Could not get image', image_id)
# pull the actual image data
self._populate_image_data(image)
return image
def add_image(self, image):
""" like set but if we already have this image from this
page we're not going to add it again. will also
fill out image stats (size, dimension) """
# we're only for new images, no i'ds allowed
# if u want to set an id by hand use set_image
if image.id:
raise o.Exception('Can not add image with id')
if not image.data:
raise o.Exception('Image must have data')
if not image.source_page_url:
raise o.Exception('Image must have source page url')
# update it's stats
image = self.populate_image_stats(image)
# only add the image if we haven't seen it beforeQ
# if we've seen it before there will be an id which
# the set of images w/ this data and from this page share
ids = self.rc.sinter('images:datainstance:%s' % image.shahash,
'images:page_ids:%s' % image.source_page_url)
# we don't need to continue
# we'll return back their original msg, w/o the id set
if ids:
print 'image already exists [%s], not setting' % ids
return image
# so the image appears to be new, good for it
return self.set_image(image)
def set_image(self, image):
""" sets image data, returns image """
# would be better if we only saved if it didn't exist
if image.data:
# save the images data
self._set_image_data(image)
# could be an update, could be new
image = self._save_to_redis(image)
# let the world know we have added a new image
self.revent.fire('image_added',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
return image
def delete_image(self, image_id):
""" removes an image """
# get it's image obj
try:
image = self.get_image(image_id)
except o.ImageNotFound, ex:
return False
# delete the redis data
self._delete_from_redis(image)
# see if we need to remove the image data
if self.rc.scard('images:datainstances:%s' % image.shahash) == 0:
# no more images w/ the same data, remove image data
with connect(Blobby) as c:
c.delete_data(image.shahash)
# it's gone, let'm know
self.revent.fire('image_deleted',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
# and we're done!
return True
def get_images_since(self, image_id=None, timestamp=None,
offset=10, limit=0):
""" returns list of tublr images or blank list which were
added after given image id or timestamp """
print '%s %s %s %s' % (image_id,timestamp,limit,offset)
if image_id is not None:
print 'got image id'
# figure out what the current id is and than grab
# our sorted set by index assuming that all ids
# contain an image
next_id = int(self.rc.get('images:next_id') or 0)
# how far from the end is the id given
d = next_id - image_id
start = next_id - d
end = next_id - d + limit - 1
print 'getting between %s %s' % (start,end)
# starting back where we think this image is to + limit
ids = self.rc.zrange('images:ids:timestamps',start,end)
print 'got ids: %s
|
pipe = self.rc.pipeline()
# if our image doesn't have an id, set it up w/ one
if not image.id:
print 'got new image: %s' % image.shahash
image.id = self.rc.incr('images:next_id')
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# check and see if we used to have a different shahash
old_shahash = self.rc.hget('images:%s' % image.id,'shahash')
if old_shahash != image.shahash:
# remove our id from the old shahash tracker
pipe.srem('images:datainstances:%s' % old_shahash,
image.id)
# add it to the new tracker
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
|
identifier_body
|
images.py
|
(object):
def __init__(self, redis_host='127.0.0.1'):
self.redis_host = redis_host
self.rc = Redis(redis_host)
self.revent = ReventClient(redis_host=self.redis_host)
# redis keys
# incr this for the next image id
# images:next_id = next_id
# all the images for the given sha
# images:datainstances:<shahash> = (ids)
# timestamp of when image was added
# images:ids:timestamps = sorted (ids,timestamp)
# all the image ids for the page
# images:page_ids:<page_url> (ids)
# last time an image was added from page
# images:pages:timestamps = sorted (url,timestamp)
# images meta data
# images:id = {}
def _image_to_dict(self, image):
data = {}
ignored_attrs = ['data']
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
if attr in ignored_attrs:
continue
v = getattr(image,attr)
if v is not None:
data[attr] = v
return data
def _dict_to_image(self, data):
image = o.Image()
for attrs in image.thrift_spec[1:]:
attr = attrs[2]
v = data.get(attr)
if v is not None:
# we might need to update the value
# type, since all values come back
# from redis as strings
attr_type = attrs[1]
# float
if attr_type == 4:
setattr(image,attr,float(v))
# int
elif attr_type == 8:
setattr(image,attr,int(v))
else:
setattr(image,attr,v)
return image
def _delete_from_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# remove it from the id set
pipe.zrem('images:ids:timestamps',image.id)
# remove it's hash
pipe.delete('images:%s' % image.id)
# decriment the count for it's image data
pipe.srem('images:datainstances:%s' % image.shahash,
image.id)
# remove image from the page's id set
if image.source_page_url:
pipe.zrem('images:page_ids:%s' % image.source_page_url,
image.id)
# make it happen
pipe.execute()
return True
def _save_to_redis(self, image):
# make these a transaction
pipe = self.rc.pipeline()
# if our image doesn't have an id, set it up w/ one
if not image.id:
print 'got new image: %s' % image.shahash
image.id = self.rc.incr('images:next_id')
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# check and see if we used to have a different shahash
old_shahash = self.rc.hget('images:%s' % image.id,'shahash')
if old_shahash != image.shahash:
# remove our id from the old shahash tracker
pipe.srem('images:datainstances:%s' % old_shahash,
image.id)
# add it to the new tracker
pipe.sadd('images:datainstances:%s' % image.shahash,
image.id)
# update / set our timestamp
da = 0.0
if image.downloaded_at:
da = image.downloaded_at
else:
da = time.time()
pipe.zadd('images:ids:timestamps',image.id, da)
# add this image to the page's id set
if image.source_page_url:
pipe.zadd('images:page_ids:%s' % image.source_page_url,
image.id, da)
# update our last scrape time for the page
pipe.zadd('images:pages:timestamps',
image.source_page_url, image.id)
# take our image and make a dict
image_data = self._image_to_dict(image)
# set our data to redis
key = 'images:%s' % image.id
pipe.hmset(key,image_data)
# execute our pipe
pipe.execute()
return image
def _get_from_redis(self, image_id):
# if the image id is in the id set than pull it's details
if self.rc.zrank('images:ids:timestamps',image_id) is not None:
# get the image data from redis
key = 'images:%s' % image_id
image_data = self.rc.hgetall(key)
if not image_data:
print 'redis had no image data'
return None
image = self._dict_to_image(image_data)
return image
return None
def _populate_image_data(self, image):
if not image.shahash:
return None
with connect(Blobby) as c:
image.data = c.get_data(image.shahash)
return image
def _set_image_data(self, image):
if image.data is not None:
with connect(Blobby) as c:
image.shahash = c.set_data(image.data)
return image
def get_image(self, image_id):
""" returns Image for given id or blank Image """
# see if we have an image
image = self._get_from_redis(image_id)
if not image:
raise o.ImageNotFound('Could not get image', image_id)
# pull the actual image data
self._populate_image_data(image)
return image
def add_image(self, image):
""" like set but if we already have this image from this
page we're not going to add it again. will also
fill out image stats (size, dimension) """
# we're only for new images, no i'ds allowed
# if u want to set an id by hand use set_image
if image.id:
raise o.Exception('Can not add image with id')
if not image.data:
raise o.Exception('Image must have data')
if not image.source_page_url:
raise o.Exception('Image must have source page url')
# update it's stats
image = self.populate_image_stats(image)
# only add the image if we haven't seen it beforeQ
# if we've seen it before there will be an id which
# the set of images w/ this data and from this page share
ids = self.rc.sinter('images:datainstance:%s' % image.shahash,
'images:page_ids:%s' % image.source_page_url)
# we don't need to continue
# we'll return back their original msg, w/o the id set
if ids:
print 'image already exists [%s], not setting' % ids
return image
# so the image appears to be new, good for it
return self.set_image(image)
def set_image(self, image):
""" sets image data, returns image """
# would be better if we only saved if it didn't exist
if image.data:
# save the images data
self._set_image_data(image)
# could be an update, could be new
image = self._save_to_redis(image)
# let the world know we have added a new image
self.revent.fire('image_added',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
return image
def delete_image(self, image_id):
""" removes an image """
# get it's image obj
try:
image = self.get_image(image_id)
except o.ImageNotFound, ex:
return False
# delete the redis data
self._delete_from_redis(image)
# see if we need to remove the image data
if self.rc.scard('images:datainstances:%s' % image.shahash) == 0:
# no more images w/ the same data, remove image data
with connect(Blobby) as c:
c.delete_data(image.shahash)
# it's gone, let'm know
self.revent.fire('image_deleted',{
'source_page_url': image.source_page_url,
'source_url': image.source_url,
'shahash': image.shahash,
'vhash': image.vhash,
'xdim': image.xdim,
'ydim': image.ydim,
})
# and we're done!
return True
def get_images_since(self, image_id=None, timestamp=None,
offset=10, limit=0):
""" returns list of tublr images or blank list which were
added after given image id or timestamp """
print '%s %s %s %s' % (image_id,timestamp,limit,offset)
if image_id is not None:
print 'got image id'
# figure out what the current id is and than grab
# our sorted set by index assuming that all ids
# contain an image
next_id = int(self.rc.get('images:next
|
ImagesHandler
|
identifier_name
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.