id
int32
0
252k
repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
75
19.8k
code_tokens
list
docstring
stringlengths
3
17.3k
docstring_tokens
list
sha
stringlengths
40
40
url
stringlengths
87
242
246,900
jmgilman/Neolib
neolib/pyamf/remoting/gateway/twisted.py
TwistedGateway._finaliseRequest
def _finaliseRequest(self, request, status, content, mimetype='text/plain'): """ Finalises the request. @param request: The HTTP Request. @type request: C{http.Request} @param status: The HTTP status code. @type status: C{int} @param content: The content of the r...
python
def _finaliseRequest(self, request, status, content, mimetype='text/plain'): """ Finalises the request. @param request: The HTTP Request. @type request: C{http.Request} @param status: The HTTP status code. @type status: C{int} @param content: The content of the r...
[ "def", "_finaliseRequest", "(", "self", ",", "request", ",", "status", ",", "content", ",", "mimetype", "=", "'text/plain'", ")", ":", "request", ".", "setResponseCode", "(", "status", ")", "request", ".", "setHeader", "(", "\"Content-Type\"", ",", "mimetype",...
Finalises the request. @param request: The HTTP Request. @type request: C{http.Request} @param status: The HTTP status code. @type status: C{int} @param content: The content of the response. @type content: C{str} @param mimetype: The MIME type of the request. ...
[ "Finalises", "the", "request", "." ]
228fafeaed0f3195676137732384a14820ae285c
https://github.com/jmgilman/Neolib/blob/228fafeaed0f3195676137732384a14820ae285c/neolib/pyamf/remoting/gateway/twisted.py#L216-L236
246,901
jmgilman/Neolib
neolib/pyamf/remoting/gateway/twisted.py
TwistedGateway.render_POST
def render_POST(self, request): """ Read remoting request from the client. @type request: The HTTP Request. @param request: C{twisted.web.http.Request} """ def handleDecodeError(failure): """ Return HTTP 400 Bad Request. """ ...
python
def render_POST(self, request): """ Read remoting request from the client. @type request: The HTTP Request. @param request: C{twisted.web.http.Request} """ def handleDecodeError(failure): """ Return HTTP 400 Bad Request. """ ...
[ "def", "render_POST", "(", "self", ",", "request", ")", ":", "def", "handleDecodeError", "(", "failure", ")", ":", "\"\"\"\n Return HTTP 400 Bad Request.\n \"\"\"", "errMesg", "=", "\"%s: %s\"", "%", "(", "failure", ".", "type", ",", "failure", ...
Read remoting request from the client. @type request: The HTTP Request. @param request: C{twisted.web.http.Request}
[ "Read", "remoting", "request", "from", "the", "client", "." ]
228fafeaed0f3195676137732384a14820ae285c
https://github.com/jmgilman/Neolib/blob/228fafeaed0f3195676137732384a14820ae285c/neolib/pyamf/remoting/gateway/twisted.py#L238-L281
246,902
kylemede/KMlogger
KMlogger/kmLogger.py
KMlogger.getStreamLevel
def getStreamLevel(self): """ Get and return current stream handler's level. """ shlvl = 0 for i in range(0,len(self.handlers)): h = self.handlers[i] if isinstance(h,logging.StreamHandler): shlvl = h.level return shlvl
python
def getStreamLevel(self): """ Get and return current stream handler's level. """ shlvl = 0 for i in range(0,len(self.handlers)): h = self.handlers[i] if isinstance(h,logging.StreamHandler): shlvl = h.level return shlvl
[ "def", "getStreamLevel", "(", "self", ")", ":", "shlvl", "=", "0", "for", "i", "in", "range", "(", "0", ",", "len", "(", "self", ".", "handlers", ")", ")", ":", "h", "=", "self", ".", "handlers", "[", "i", "]", "if", "isinstance", "(", "h", ","...
Get and return current stream handler's level.
[ "Get", "and", "return", "current", "stream", "handler", "s", "level", "." ]
c904f0aaf35b4aff259567f932b5ca678d7f8176
https://github.com/kylemede/KMlogger/blob/c904f0aaf35b4aff259567f932b5ca678d7f8176/KMlogger/kmLogger.py#L89-L98
246,903
kylemede/KMlogger
KMlogger/kmLogger.py
KMlogger.logSystemInfo
def logSystemInfo(self): """ A function to be called just after a logging object is instantiated to load the log up with info about the computer it is being ran on and the software version. This function utilizes the psutil and platform libraries, so they must be install for...
python
def logSystemInfo(self): """ A function to be called just after a logging object is instantiated to load the log up with info about the computer it is being ran on and the software version. This function utilizes the psutil and platform libraries, so they must be install for...
[ "def", "logSystemInfo", "(", "self", ")", ":", "t", "=", "datetime", ".", "date", ".", "today", "(", ")", "infoStr", "=", "'Date KMlogger object instantiated: '", "+", "t", ".", "strftime", "(", "'%b %d, %Y'", ")", "+", "'\\n\\n'", "infoStr", "+=", "\"\\n\""...
A function to be called just after a logging object is instantiated to load the log up with info about the computer it is being ran on and the software version. This function utilizes the psutil and platform libraries, so they must be install for it to work. For clarity of the log,...
[ "A", "function", "to", "be", "called", "just", "after", "a", "logging", "object", "is", "instantiated", "to", "load", "the", "log", "up", "with", "info", "about", "the", "computer", "it", "is", "being", "ran", "on", "and", "the", "software", "version", "...
c904f0aaf35b4aff259567f932b5ca678d7f8176
https://github.com/kylemede/KMlogger/blob/c904f0aaf35b4aff259567f932b5ca678d7f8176/KMlogger/kmLogger.py#L137-L173
246,904
exekias/droplet
droplet/module.py
ModuleMeta.install_wrapper
def install_wrapper(cls, install, new_class): """ Wrap the install method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if self.installed: raise AssertionError('Module %s is already installed' ...
python
def install_wrapper(cls, install, new_class): """ Wrap the install method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if self.installed: raise AssertionError('Module %s is already installed' ...
[ "def", "install_wrapper", "(", "cls", ",", "install", ",", "new_class", ")", ":", "def", "_wrapped", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "self", ".", "installed", ":", "raise", "AssertionError", "(", "'Module %s is alr...
Wrap the install method to call pre and post enable signals and update module status
[ "Wrap", "the", "install", "method", "to", "call", "pre", "and", "post", "enable", "signals", "and", "update", "module", "status" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L70-L92
246,905
exekias/droplet
droplet/module.py
ModuleMeta.enable_wrapper
def enable_wrapper(cls, enable, new_class): """ Wrap the enable method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.installed: raise AssertionError('Module %s cannot be enabled' ...
python
def enable_wrapper(cls, enable, new_class): """ Wrap the enable method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.installed: raise AssertionError('Module %s cannot be enabled' ...
[ "def", "enable_wrapper", "(", "cls", ",", "enable", ",", "new_class", ")", ":", "def", "_wrapped", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "not", "self", ".", "installed", ":", "raise", "AssertionError", "(", "'Module %s...
Wrap the enable method to call pre and post enable signals and update module status
[ "Wrap", "the", "enable", "method", "to", "call", "pre", "and", "post", "enable", "signals", "and", "update", "module", "status" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L95-L127
246,906
exekias/droplet
droplet/module.py
ModuleMeta.save_wrapper
def save_wrapper(cls, save, new_class): """ Wrap the save method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.installed: raise AssertionError('Module %s is not installed' % ...
python
def save_wrapper(cls, save, new_class): """ Wrap the save method to call pre and post enable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.installed: raise AssertionError('Module %s is not installed' % ...
[ "def", "save_wrapper", "(", "cls", ",", "save", ",", "new_class", ")", ":", "def", "_wrapped", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "not", "self", ".", "installed", ":", "raise", "AssertionError", "(", "'Module %s is ...
Wrap the save method to call pre and post enable signals and update module status
[ "Wrap", "the", "save", "method", "to", "call", "pre", "and", "post", "enable", "signals", "and", "update", "module", "status" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L130-L148
246,907
exekias/droplet
droplet/module.py
ModuleMeta.disable_wrapper
def disable_wrapper(cls, disable, new_class): """ Wrap the disable method to call pre and post disable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.enabled: raise AssertionError('Module %s is already disabled' ...
python
def disable_wrapper(cls, disable, new_class): """ Wrap the disable method to call pre and post disable signals and update module status """ def _wrapped(self, *args, **kwargs): if not self.enabled: raise AssertionError('Module %s is already disabled' ...
[ "def", "disable_wrapper", "(", "cls", ",", "disable", ",", "new_class", ")", ":", "def", "_wrapped", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "not", "self", ".", "enabled", ":", "raise", "AssertionError", "(", "'Module %s...
Wrap the disable method to call pre and post disable signals and update module status
[ "Wrap", "the", "disable", "method", "to", "call", "pre", "and", "post", "disable", "signals", "and", "update", "module", "status" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L151-L178
246,908
exekias/droplet
droplet/module.py
Module._info
def _info(self): """ Module internal status representation """ name = self.__class__.__module__ + '.' + self.__class__.__name__ info, created = ModuleInfo.objects.get_or_create(name=name) if created: # Do not set as changed info.commit() re...
python
def _info(self): """ Module internal status representation """ name = self.__class__.__module__ + '.' + self.__class__.__name__ info, created = ModuleInfo.objects.get_or_create(name=name) if created: # Do not set as changed info.commit() re...
[ "def", "_info", "(", "self", ")", ":", "name", "=", "self", ".", "__class__", ".", "__module__", "+", "'.'", "+", "self", ".", "__class__", ".", "__name__", "info", ",", "created", "=", "ModuleInfo", ".", "objects", ".", "get_or_create", "(", "name", "...
Module internal status representation
[ "Module", "internal", "status", "representation" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L266-L275
246,909
exekias/droplet
droplet/module.py
Module.models
def models(self): """ Return all the models defined for this module """ app = get_app(self.__class__.__module__.split('.')[-2]) return get_models(app)
python
def models(self): """ Return all the models defined for this module """ app = get_app(self.__class__.__module__.split('.')[-2]) return get_models(app)
[ "def", "models", "(", "self", ")", ":", "app", "=", "get_app", "(", "self", ".", "__class__", ".", "__module__", ".", "split", "(", "'.'", ")", "[", "-", "2", "]", ")", "return", "get_models", "(", "app", ")" ]
Return all the models defined for this module
[ "Return", "all", "the", "models", "defined", "for", "this", "module" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L338-L343
246,910
exekias/droplet
droplet/module.py
Module.conf_files
def conf_files(self): """ List of configuration files for this module """ for attr in dir(self): field = getattr(self, attr) if isinstance(field, ConfFile): yield field
python
def conf_files(self): """ List of configuration files for this module """ for attr in dir(self): field = getattr(self, attr) if isinstance(field, ConfFile): yield field
[ "def", "conf_files", "(", "self", ")", ":", "for", "attr", "in", "dir", "(", "self", ")", ":", "field", "=", "getattr", "(", "self", ",", "attr", ")", "if", "isinstance", "(", "field", ",", "ConfFile", ")", ":", "yield", "field" ]
List of configuration files for this module
[ "List", "of", "configuration", "files", "for", "this", "module" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L380-L387
246,911
exekias/droplet
droplet/module.py
Module.daemons
def daemons(self): """ List of daemons for this module """ for attr in dir(self): field = getattr(self, attr) if isinstance(field, Daemon): yield field
python
def daemons(self): """ List of daemons for this module """ for attr in dir(self): field = getattr(self, attr) if isinstance(field, Daemon): yield field
[ "def", "daemons", "(", "self", ")", ":", "for", "attr", "in", "dir", "(", "self", ")", ":", "field", "=", "getattr", "(", "self", ",", "attr", ")", "if", "isinstance", "(", "field", ",", "Daemon", ")", ":", "yield", "field" ]
List of daemons for this module
[ "List", "of", "daemons", "for", "this", "module" ]
aeac573a2c1c4b774e99d5414a1c79b1bb734941
https://github.com/exekias/droplet/blob/aeac573a2c1c4b774e99d5414a1c79b1bb734941/droplet/module.py#L419-L426
246,912
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_endpoint_by_endpoint_id
def get_endpoint_by_endpoint_id(self, endpoint_id): """ Get an endpoint by endpoint id """ self._validate_uuid(endpoint_id) url = "/notification/v1/endpoint/{}".format(endpoint_id) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: ...
python
def get_endpoint_by_endpoint_id(self, endpoint_id): """ Get an endpoint by endpoint id """ self._validate_uuid(endpoint_id) url = "/notification/v1/endpoint/{}".format(endpoint_id) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: ...
[ "def", "get_endpoint_by_endpoint_id", "(", "self", ",", "endpoint_id", ")", ":", "self", ".", "_validate_uuid", "(", "endpoint_id", ")", "url", "=", "\"/notification/v1/endpoint/{}\"", ".", "format", "(", "endpoint_id", ")", "response", "=", "NWS_DAO", "(", ")", ...
Get an endpoint by endpoint id
[ "Get", "an", "endpoint", "by", "endpoint", "id" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L46-L59
246,913
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_endpoint_by_subscriber_id_and_protocol
def get_endpoint_by_subscriber_id_and_protocol( self, subscriber_id, protocol): """ Get an endpoint by subscriber_id and protocol """ self._validate_subscriber_id(subscriber_id) self._validate_endpoint_protocol(protocol) url = "/notification/v1/endpoint?subsc...
python
def get_endpoint_by_subscriber_id_and_protocol( self, subscriber_id, protocol): """ Get an endpoint by subscriber_id and protocol """ self._validate_subscriber_id(subscriber_id) self._validate_endpoint_protocol(protocol) url = "/notification/v1/endpoint?subsc...
[ "def", "get_endpoint_by_subscriber_id_and_protocol", "(", "self", ",", "subscriber_id", ",", "protocol", ")", ":", "self", ".", "_validate_subscriber_id", "(", "subscriber_id", ")", "self", ".", "_validate_endpoint_protocol", "(", "protocol", ")", "url", "=", "\"/noti...
Get an endpoint by subscriber_id and protocol
[ "Get", "an", "endpoint", "by", "subscriber_id", "and", "protocol" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L61-L81
246,914
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_endpoint_by_address
def get_endpoint_by_address(self, endpoint_addr): """ Get an endpoint by address """ url = "/notification/v1/endpoint?endpoint_address={}".format( endpoint_addr) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: rais...
python
def get_endpoint_by_address(self, endpoint_addr): """ Get an endpoint by address """ url = "/notification/v1/endpoint?endpoint_address={}".format( endpoint_addr) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: rais...
[ "def", "get_endpoint_by_address", "(", "self", ",", "endpoint_addr", ")", ":", "url", "=", "\"/notification/v1/endpoint?endpoint_address={}\"", ".", "format", "(", "endpoint_addr", ")", "response", "=", "NWS_DAO", "(", ")", ".", "getURL", "(", "url", ",", "self", ...
Get an endpoint by address
[ "Get", "an", "endpoint", "by", "address" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L83-L99
246,915
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_endpoints_by_subscriber_id
def get_endpoints_by_subscriber_id(self, subscriber_id): """ Search for all endpoints by a given subscriber """ self._validate_subscriber_id(subscriber_id) url = "/notification/v1/endpoint?subscriber_id={}".format( subscriber_id) response = NWS_DAO().getURL(...
python
def get_endpoints_by_subscriber_id(self, subscriber_id): """ Search for all endpoints by a given subscriber """ self._validate_subscriber_id(subscriber_id) url = "/notification/v1/endpoint?subscriber_id={}".format( subscriber_id) response = NWS_DAO().getURL(...
[ "def", "get_endpoints_by_subscriber_id", "(", "self", ",", "subscriber_id", ")", ":", "self", ".", "_validate_subscriber_id", "(", "subscriber_id", ")", "url", "=", "\"/notification/v1/endpoint?subscriber_id={}\"", ".", "format", "(", "subscriber_id", ")", "response", "...
Search for all endpoints by a given subscriber
[ "Search", "for", "all", "endpoints", "by", "a", "given", "subscriber" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L101-L120
246,916
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.resend_sms_endpoint_verification
def resend_sms_endpoint_verification(self, endpoint_id): """ Calls NWS function to resend verification message to endpoint's phone number """ self._validate_uuid(endpoint_id) url = "/notification/v1/endpoint/{}/verification".format(endpoint_id) response = NWS_DA...
python
def resend_sms_endpoint_verification(self, endpoint_id): """ Calls NWS function to resend verification message to endpoint's phone number """ self._validate_uuid(endpoint_id) url = "/notification/v1/endpoint/{}/verification".format(endpoint_id) response = NWS_DA...
[ "def", "resend_sms_endpoint_verification", "(", "self", ",", "endpoint_id", ")", ":", "self", ".", "_validate_uuid", "(", "endpoint_id", ")", "url", "=", "\"/notification/v1/endpoint/{}/verification\"", ".", "format", "(", "endpoint_id", ")", "response", "=", "NWS_DAO...
Calls NWS function to resend verification message to endpoint's phone number
[ "Calls", "NWS", "function", "to", "resend", "verification", "message", "to", "endpoint", "s", "phone", "number" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L122-L135
246,917
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_subscriptions_by_subscriber_id
def get_subscriptions_by_subscriber_id( self, subscriber_id, max_results=10): """ Search for all subscriptions by a given subscriber """ return self.search_subscriptions( subscriber_id=subscriber_id, max_results=max_results)
python
def get_subscriptions_by_subscriber_id( self, subscriber_id, max_results=10): """ Search for all subscriptions by a given subscriber """ return self.search_subscriptions( subscriber_id=subscriber_id, max_results=max_results)
[ "def", "get_subscriptions_by_subscriber_id", "(", "self", ",", "subscriber_id", ",", "max_results", "=", "10", ")", ":", "return", "self", ".", "search_subscriptions", "(", "subscriber_id", "=", "subscriber_id", ",", "max_results", "=", "max_results", ")" ]
Search for all subscriptions by a given subscriber
[ "Search", "for", "all", "subscriptions", "by", "a", "given", "subscriber" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L237-L243
246,918
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_subscriptions_by_channel_id_and_subscriber_id
def get_subscriptions_by_channel_id_and_subscriber_id( self, channel_id, subscriber_id): """ Search for all subscriptions by a given channel and subscriber """ return self.search_subscriptions( channel_id=channel_id, subscriber_id=subscriber_id)
python
def get_subscriptions_by_channel_id_and_subscriber_id( self, channel_id, subscriber_id): """ Search for all subscriptions by a given channel and subscriber """ return self.search_subscriptions( channel_id=channel_id, subscriber_id=subscriber_id)
[ "def", "get_subscriptions_by_channel_id_and_subscriber_id", "(", "self", ",", "channel_id", ",", "subscriber_id", ")", ":", "return", "self", ".", "search_subscriptions", "(", "channel_id", "=", "channel_id", ",", "subscriber_id", "=", "subscriber_id", ")" ]
Search for all subscriptions by a given channel and subscriber
[ "Search", "for", "all", "subscriptions", "by", "a", "given", "channel", "and", "subscriber" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L245-L251
246,919
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_subscriptions_by_channel_id_and_person_id
def get_subscriptions_by_channel_id_and_person_id( self, channel_id, person_id): """ Search for all subscriptions by a given channel and person """ return self.search_subscriptions( channel_id=channel_id, person_id=person_id)
python
def get_subscriptions_by_channel_id_and_person_id( self, channel_id, person_id): """ Search for all subscriptions by a given channel and person """ return self.search_subscriptions( channel_id=channel_id, person_id=person_id)
[ "def", "get_subscriptions_by_channel_id_and_person_id", "(", "self", ",", "channel_id", ",", "person_id", ")", ":", "return", "self", ".", "search_subscriptions", "(", "channel_id", "=", "channel_id", ",", "person_id", "=", "person_id", ")" ]
Search for all subscriptions by a given channel and person
[ "Search", "for", "all", "subscriptions", "by", "a", "given", "channel", "and", "person" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L253-L259
246,920
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_subscription_by_channel_id_and_endpoint_id
def get_subscription_by_channel_id_and_endpoint_id( self, channel_id, endpoint_id): """ Search for subscription by a given channel and endpoint """ subscriptions = self.search_subscriptions( channel_id=channel_id, endpoint_id=endpoint_id) try: ...
python
def get_subscription_by_channel_id_and_endpoint_id( self, channel_id, endpoint_id): """ Search for subscription by a given channel and endpoint """ subscriptions = self.search_subscriptions( channel_id=channel_id, endpoint_id=endpoint_id) try: ...
[ "def", "get_subscription_by_channel_id_and_endpoint_id", "(", "self", ",", "channel_id", ",", "endpoint_id", ")", ":", "subscriptions", "=", "self", ".", "search_subscriptions", "(", "channel_id", "=", "channel_id", ",", "endpoint_id", "=", "endpoint_id", ")", "try", ...
Search for subscription by a given channel and endpoint
[ "Search", "for", "subscription", "by", "a", "given", "channel", "and", "endpoint" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L261-L272
246,921
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.search_subscriptions
def search_subscriptions(self, **kwargs): """ Search for all subscriptions by parameters """ params = [(key, kwargs[key]) for key in sorted(kwargs.keys())] url = "/notification/v1/subscription?{}".format( urlencode(params, doseq=True)) response = NWS_DAO().ge...
python
def search_subscriptions(self, **kwargs): """ Search for all subscriptions by parameters """ params = [(key, kwargs[key]) for key in sorted(kwargs.keys())] url = "/notification/v1/subscription?{}".format( urlencode(params, doseq=True)) response = NWS_DAO().ge...
[ "def", "search_subscriptions", "(", "self", ",", "*", "*", "kwargs", ")", ":", "params", "=", "[", "(", "key", ",", "kwargs", "[", "key", "]", ")", "for", "key", "in", "sorted", "(", "kwargs", ".", "keys", "(", ")", ")", "]", "url", "=", "\"/noti...
Search for all subscriptions by parameters
[ "Search", "for", "all", "subscriptions", "by", "parameters" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L274-L291
246,922
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_channel_by_channel_id
def get_channel_by_channel_id(self, channel_id): """ Get a channel by channel id """ self._validate_uuid(channel_id) url = "/notification/v1/channel/{}".format(channel_id) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: ...
python
def get_channel_by_channel_id(self, channel_id): """ Get a channel by channel id """ self._validate_uuid(channel_id) url = "/notification/v1/channel/{}".format(channel_id) response = NWS_DAO().getURL(url, self._read_headers) if response.status != 200: ...
[ "def", "get_channel_by_channel_id", "(", "self", ",", "channel_id", ")", ":", "self", ".", "_validate_uuid", "(", "channel_id", ")", "url", "=", "\"/notification/v1/channel/{}\"", ".", "format", "(", "channel_id", ")", "response", "=", "NWS_DAO", "(", ")", ".", ...
Get a channel by channel id
[ "Get", "a", "channel", "by", "channel", "id" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L293-L307
246,923
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_channels_by_sln
def get_channels_by_sln(self, channel_type, sln): """ Search for all channels by sln """ return self.search_channels(type=channel_type, tag_sln=sln)
python
def get_channels_by_sln(self, channel_type, sln): """ Search for all channels by sln """ return self.search_channels(type=channel_type, tag_sln=sln)
[ "def", "get_channels_by_sln", "(", "self", ",", "channel_type", ",", "sln", ")", ":", "return", "self", ".", "search_channels", "(", "type", "=", "channel_type", ",", "tag_sln", "=", "sln", ")" ]
Search for all channels by sln
[ "Search", "for", "all", "channels", "by", "sln" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L309-L313
246,924
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_channels_by_sln_year_quarter
def get_channels_by_sln_year_quarter( self, channel_type, sln, year, quarter): """ Search for all channels by sln, year and quarter """ return self.search_channels( type=channel_type, tag_sln=sln, tag_year=year, tag_quarter=quarter)
python
def get_channels_by_sln_year_quarter( self, channel_type, sln, year, quarter): """ Search for all channels by sln, year and quarter """ return self.search_channels( type=channel_type, tag_sln=sln, tag_year=year, tag_quarter=quarter)
[ "def", "get_channels_by_sln_year_quarter", "(", "self", ",", "channel_type", ",", "sln", ",", "year", ",", "quarter", ")", ":", "return", "self", ".", "search_channels", "(", "type", "=", "channel_type", ",", "tag_sln", "=", "sln", ",", "tag_year", "=", "yea...
Search for all channels by sln, year and quarter
[ "Search", "for", "all", "channels", "by", "sln", "year", "and", "quarter" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L315-L321
246,925
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.get_active_channels_by_year_quarter
def get_active_channels_by_year_quarter( self, channel_type, year, quarter, expires=None): """ Search for all active channels by year and quarter """ if expires is None: # Set expires_after to midnight of current day expires = datetime.combine(datetime...
python
def get_active_channels_by_year_quarter( self, channel_type, year, quarter, expires=None): """ Search for all active channels by year and quarter """ if expires is None: # Set expires_after to midnight of current day expires = datetime.combine(datetime...
[ "def", "get_active_channels_by_year_quarter", "(", "self", ",", "channel_type", ",", "year", ",", "quarter", ",", "expires", "=", "None", ")", ":", "if", "expires", "is", "None", ":", "# Set expires_after to midnight of current day", "expires", "=", "datetime", ".",...
Search for all active channels by year and quarter
[ "Search", "for", "all", "active", "channels", "by", "year", "and", "quarter" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L323-L334
246,926
uw-it-aca/uw-restclients-nws
uw_nws/__init__.py
NWS.search_channels
def search_channels(self, **kwargs): """ Search for all channels by parameters """ params = [(key, kwargs[key]) for key in sorted(kwargs.keys())] url = "/notification/v1/channel?{}".format( urlencode(params, doseq=True)) response = NWS_DAO().getURL(url, self....
python
def search_channels(self, **kwargs): """ Search for all channels by parameters """ params = [(key, kwargs[key]) for key in sorted(kwargs.keys())] url = "/notification/v1/channel?{}".format( urlencode(params, doseq=True)) response = NWS_DAO().getURL(url, self....
[ "def", "search_channels", "(", "self", ",", "*", "*", "kwargs", ")", ":", "params", "=", "[", "(", "key", ",", "kwargs", "[", "key", "]", ")", "for", "key", "in", "sorted", "(", "kwargs", ".", "keys", "(", ")", ")", "]", "url", "=", "\"/notificat...
Search for all channels by parameters
[ "Search", "for", "all", "channels", "by", "parameters" ]
ec6fd14342ffc883d14bcb53b2fe9bc288696027
https://github.com/uw-it-aca/uw-restclients-nws/blob/ec6fd14342ffc883d14bcb53b2fe9bc288696027/uw_nws/__init__.py#L336-L353
246,927
knagra/farnsworth
base/pipeline.py
_get_first_last
def _get_first_last(details): """ Gets a user's first and last name from details. """ if "first_name" in details and "last_name" in details: return details["first_name"], details["last_name"] elif "first_name" in details: lst = details["first_name"].rsplit(" ", 1) if len(lst)...
python
def _get_first_last(details): """ Gets a user's first and last name from details. """ if "first_name" in details and "last_name" in details: return details["first_name"], details["last_name"] elif "first_name" in details: lst = details["first_name"].rsplit(" ", 1) if len(lst)...
[ "def", "_get_first_last", "(", "details", ")", ":", "if", "\"first_name\"", "in", "details", "and", "\"last_name\"", "in", "details", ":", "return", "details", "[", "\"first_name\"", "]", ",", "details", "[", "\"last_name\"", "]", "elif", "\"first_name\"", "in",...
Gets a user's first and last name from details.
[ "Gets", "a", "user", "s", "first", "and", "last", "name", "from", "details", "." ]
1b6589f0d9fea154f0a1e2231ed906764ed26d26
https://github.com/knagra/farnsworth/blob/1b6589f0d9fea154f0a1e2231ed906764ed26d26/base/pipeline.py#L17-L32
246,928
minhhoit/yacms
yacms/boot/__init__.py
import_field
def import_field(field_classpath): """ Imports a field by its dotted class path, prepending "django.db.models" to raw class names and raising an exception if the import fails. """ if '.' in field_classpath: fully_qualified = field_classpath else: fully_qualified = "django.db.mode...
python
def import_field(field_classpath): """ Imports a field by its dotted class path, prepending "django.db.models" to raw class names and raising an exception if the import fails. """ if '.' in field_classpath: fully_qualified = field_classpath else: fully_qualified = "django.db.mode...
[ "def", "import_field", "(", "field_classpath", ")", ":", "if", "'.'", "in", "field_classpath", ":", "fully_qualified", "=", "field_classpath", "else", ":", "fully_qualified", "=", "\"django.db.models.%s\"", "%", "field_classpath", "try", ":", "return", "import_dotted_...
Imports a field by its dotted class path, prepending "django.db.models" to raw class names and raising an exception if the import fails.
[ "Imports", "a", "field", "by", "its", "dotted", "class", "path", "prepending", "django", ".", "db", ".", "models", "to", "raw", "class", "names", "and", "raising", "an", "exception", "if", "the", "import", "fails", "." ]
2921b706b7107c6e8c5f2bbf790ff11f85a2167f
https://github.com/minhhoit/yacms/blob/2921b706b7107c6e8c5f2bbf790ff11f85a2167f/yacms/boot/__init__.py#L38-L52
246,929
minhhoit/yacms
yacms/boot/__init__.py
add_extra_model_fields
def add_extra_model_fields(sender, **kwargs): """ Injects custom fields onto the given sender model as defined by the ``EXTRA_MODEL_FIELDS`` setting. This is a closure over the "fields" variable. """ model_key = sender._meta.app_label, sender._meta.model_name for field_name, field in fields....
python
def add_extra_model_fields(sender, **kwargs): """ Injects custom fields onto the given sender model as defined by the ``EXTRA_MODEL_FIELDS`` setting. This is a closure over the "fields" variable. """ model_key = sender._meta.app_label, sender._meta.model_name for field_name, field in fields....
[ "def", "add_extra_model_fields", "(", "sender", ",", "*", "*", "kwargs", ")", ":", "model_key", "=", "sender", ".", "_meta", ".", "app_label", ",", "sender", ".", "_meta", ".", "model_name", "for", "field_name", ",", "field", "in", "fields", ".", "get", ...
Injects custom fields onto the given sender model as defined by the ``EXTRA_MODEL_FIELDS`` setting. This is a closure over the "fields" variable.
[ "Injects", "custom", "fields", "onto", "the", "given", "sender", "model", "as", "defined", "by", "the", "EXTRA_MODEL_FIELDS", "setting", ".", "This", "is", "a", "closure", "over", "the", "fields", "variable", "." ]
2921b706b7107c6e8c5f2bbf790ff11f85a2167f
https://github.com/minhhoit/yacms/blob/2921b706b7107c6e8c5f2bbf790ff11f85a2167f/yacms/boot/__init__.py#L80-L88
246,930
jmvrbanac/PackMap
packmap/finder.py
main
def main(*args, **kwargs): """ Simple entry-point that takes the package name and json output""" parser = argparse.ArgumentParser() parser.add_argument('package_name', type=str) parser.add_argument('output_filename', type=str) parsed_args = parser.parse_args() # Find all package requirements ...
python
def main(*args, **kwargs): """ Simple entry-point that takes the package name and json output""" parser = argparse.ArgumentParser() parser.add_argument('package_name', type=str) parser.add_argument('output_filename', type=str) parsed_args = parser.parse_args() # Find all package requirements ...
[ "def", "main", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", ")", "parser", ".", "add_argument", "(", "'package_name'", ",", "type", "=", "str", ")", "parser", ".", "add_argument", "(", "'ou...
Simple entry-point that takes the package name and json output
[ "Simple", "entry", "-", "point", "that", "takes", "the", "package", "name", "and", "json", "output" ]
e35d12d21ab109cae1175e0dc94d7f4855256b23
https://github.com/jmvrbanac/PackMap/blob/e35d12d21ab109cae1175e0dc94d7f4855256b23/packmap/finder.py#L57-L71
246,931
wdbm/shijian
shijian.py
tail
def tail( filepath = "log.txt", lines = 50 ): """ Return a specified number of last lines of a specified file. If there is an error or the file does not exist, return False. """ try: filepath = os.path.expanduser(os.path.expandvars(filepath)) if os.path.isfile(filepath...
python
def tail( filepath = "log.txt", lines = 50 ): """ Return a specified number of last lines of a specified file. If there is an error or the file does not exist, return False. """ try: filepath = os.path.expanduser(os.path.expandvars(filepath)) if os.path.isfile(filepath...
[ "def", "tail", "(", "filepath", "=", "\"log.txt\"", ",", "lines", "=", "50", ")", ":", "try", ":", "filepath", "=", "os", ".", "path", ".", "expanduser", "(", "os", ".", "path", ".", "expandvars", "(", "filepath", ")", ")", "if", "os", ".", "path",...
Return a specified number of last lines of a specified file. If there is an error or the file does not exist, return False.
[ "Return", "a", "specified", "number", "of", "last", "lines", "of", "a", "specified", "file", ".", "If", "there", "is", "an", "error", "or", "the", "file", "does", "not", "exist", "return", "False", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L657-L676
246,932
wdbm/shijian
shijian.py
convert_type_list_elements
def convert_type_list_elements( list_object = None, element_type = str ): """ Recursively convert all elements and all elements of all sublists of a list to a specified type and return the new list. """ if element_type is str: return [str(element) if not isinstance(element, list...
python
def convert_type_list_elements( list_object = None, element_type = str ): """ Recursively convert all elements and all elements of all sublists of a list to a specified type and return the new list. """ if element_type is str: return [str(element) if not isinstance(element, list...
[ "def", "convert_type_list_elements", "(", "list_object", "=", "None", ",", "element_type", "=", "str", ")", ":", "if", "element_type", "is", "str", ":", "return", "[", "str", "(", "element", ")", "if", "not", "isinstance", "(", "element", ",", "list", ")",...
Recursively convert all elements and all elements of all sublists of a list to a specified type and return the new list.
[ "Recursively", "convert", "all", "elements", "and", "all", "elements", "of", "all", "sublists", "of", "a", "list", "to", "a", "specified", "type", "and", "return", "the", "new", "list", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L878-L890
246,933
wdbm/shijian
shijian.py
select_spread
def select_spread( list_of_elements = None, number_of_elements = None ): """ This function returns the specified number of elements of a list spread approximately evenly. """ if len(list_of_elements) <= number_of_elements: return list_of_elements if number_of_elements == 0:...
python
def select_spread( list_of_elements = None, number_of_elements = None ): """ This function returns the specified number of elements of a list spread approximately evenly. """ if len(list_of_elements) <= number_of_elements: return list_of_elements if number_of_elements == 0:...
[ "def", "select_spread", "(", "list_of_elements", "=", "None", ",", "number_of_elements", "=", "None", ")", ":", "if", "len", "(", "list_of_elements", ")", "<=", "number_of_elements", ":", "return", "list_of_elements", "if", "number_of_elements", "==", "0", ":", ...
This function returns the specified number of elements of a list spread approximately evenly.
[ "This", "function", "returns", "the", "specified", "number", "of", "elements", "of", "a", "list", "spread", "approximately", "evenly", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L997-L1015
246,934
wdbm/shijian
shijian.py
split_list
def split_list( list_object = None, granularity = None ): """ This function splits a list into a specified number of lists. It returns a list of lists that correspond to these parts. Negative numbers of parts are not accepted and numbers of parts greater than the number of elements in the ...
python
def split_list( list_object = None, granularity = None ): """ This function splits a list into a specified number of lists. It returns a list of lists that correspond to these parts. Negative numbers of parts are not accepted and numbers of parts greater than the number of elements in the ...
[ "def", "split_list", "(", "list_object", "=", "None", ",", "granularity", "=", "None", ")", ":", "if", "granularity", "<", "0", ":", "raise", "Exception", "(", "\"negative granularity\"", ")", "mean_length", "=", "len", "(", "list_object", ")", "/", "float",...
This function splits a list into a specified number of lists. It returns a list of lists that correspond to these parts. Negative numbers of parts are not accepted and numbers of parts greater than the number of elements in the list result in the maximum possible number of lists being returned.
[ "This", "function", "splits", "a", "list", "into", "a", "specified", "number", "of", "lists", ".", "It", "returns", "a", "list", "of", "lists", "that", "correspond", "to", "these", "parts", ".", "Negative", "numbers", "of", "parts", "are", "not", "accepted...
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1017-L1040
246,935
wdbm/shijian
shijian.py
ustr
def ustr(text): """ Convert a string to Python 2 unicode or Python 3 string as appropriate to the version of Python in use. """ if text is not None: if sys.version_info >= (3, 0): return str(text) else: return unicode(text) else: return text
python
def ustr(text): """ Convert a string to Python 2 unicode or Python 3 string as appropriate to the version of Python in use. """ if text is not None: if sys.version_info >= (3, 0): return str(text) else: return unicode(text) else: return text
[ "def", "ustr", "(", "text", ")", ":", "if", "text", "is", "not", "None", ":", "if", "sys", ".", "version_info", ">=", "(", "3", ",", "0", ")", ":", "return", "str", "(", "text", ")", "else", ":", "return", "unicode", "(", "text", ")", "else", "...
Convert a string to Python 2 unicode or Python 3 string as appropriate to the version of Python in use.
[ "Convert", "a", "string", "to", "Python", "2", "unicode", "or", "Python", "3", "string", "as", "appropriate", "to", "the", "version", "of", "Python", "in", "use", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1231-L1242
246,936
wdbm/shijian
shijian.py
add_time_variables
def add_time_variables(df, reindex = True): """ Return a DataFrame with variables for weekday index, weekday name, timedelta through day, fraction through day, hours through day and days through week added, optionally with the index set to datetime and the variable `datetime` removed. It is assumed ...
python
def add_time_variables(df, reindex = True): """ Return a DataFrame with variables for weekday index, weekday name, timedelta through day, fraction through day, hours through day and days through week added, optionally with the index set to datetime and the variable `datetime` removed. It is assumed ...
[ "def", "add_time_variables", "(", "df", ",", "reindex", "=", "True", ")", ":", "if", "not", "\"datetime\"", "in", "df", ".", "columns", ":", "log", ".", "error", "(", "\"field datetime not found in DataFrame\"", ")", "return", "False", "df", "[", "\"datetime\"...
Return a DataFrame with variables for weekday index, weekday name, timedelta through day, fraction through day, hours through day and days through week added, optionally with the index set to datetime and the variable `datetime` removed. It is assumed that the variable `datetime` exists.
[ "Return", "a", "DataFrame", "with", "variables", "for", "weekday", "index", "weekday", "name", "timedelta", "through", "day", "fraction", "through", "day", "hours", "through", "day", "and", "days", "through", "week", "added", "optionally", "with", "the", "index"...
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1578-L1611
246,937
wdbm/shijian
shijian.py
daily_plots
def daily_plots( df, variable, renormalize = True, plot = True, scatter = False, linestyle = "-", linewidth = 1, s = 1 ): """ Create daily plots of a variable in a DataFrame, optionally renormalized. It is assumed that the DataFrame index is datet...
python
def daily_plots( df, variable, renormalize = True, plot = True, scatter = False, linestyle = "-", linewidth = 1, s = 1 ): """ Create daily plots of a variable in a DataFrame, optionally renormalized. It is assumed that the DataFrame index is datet...
[ "def", "daily_plots", "(", "df", ",", "variable", ",", "renormalize", "=", "True", ",", "plot", "=", "True", ",", "scatter", "=", "False", ",", "linestyle", "=", "\"-\"", ",", "linewidth", "=", "1", ",", "s", "=", "1", ")", ":", "if", "not", "df", ...
Create daily plots of a variable in a DataFrame, optionally renormalized. It is assumed that the DataFrame index is datetime.
[ "Create", "daily", "plots", "of", "a", "variable", "in", "a", "DataFrame", "optionally", "renormalized", ".", "It", "is", "assumed", "that", "the", "DataFrame", "index", "is", "datetime", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1613-L1644
246,938
wdbm/shijian
shijian.py
weekly_plots
def weekly_plots( df, variable, renormalize = True, plot = True, scatter = False, linestyle = "-", linewidth = 1, s = 1 ): """ Create weekly plots of a variable in a DataFrame, optionally renormalized. It is assumed that the variable `days_through...
python
def weekly_plots( df, variable, renormalize = True, plot = True, scatter = False, linestyle = "-", linewidth = 1, s = 1 ): """ Create weekly plots of a variable in a DataFrame, optionally renormalized. It is assumed that the variable `days_through...
[ "def", "weekly_plots", "(", "df", ",", "variable", ",", "renormalize", "=", "True", ",", "plot", "=", "True", ",", "scatter", "=", "False", ",", "linestyle", "=", "\"-\"", ",", "linewidth", "=", "1", ",", "s", "=", "1", ")", ":", "if", "not", "\"da...
Create weekly plots of a variable in a DataFrame, optionally renormalized. It is assumed that the variable `days_through_week` exists.
[ "Create", "weekly", "plots", "of", "a", "variable", "in", "a", "DataFrame", "optionally", "renormalized", ".", "It", "is", "assumed", "that", "the", "variable", "days_through_week", "exists", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1646-L1680
246,939
wdbm/shijian
shijian.py
yearly_plots
def yearly_plots( df, variable, renormalize = True, horizontal_axis_labels_days = False, horizontal_axis_labels_months = True, plot = True, scatter = False, linestyle = "-", linewidth ...
python
def yearly_plots( df, variable, renormalize = True, horizontal_axis_labels_days = False, horizontal_axis_labels_months = True, plot = True, scatter = False, linestyle = "-", linewidth ...
[ "def", "yearly_plots", "(", "df", ",", "variable", ",", "renormalize", "=", "True", ",", "horizontal_axis_labels_days", "=", "False", ",", "horizontal_axis_labels_months", "=", "True", ",", "plot", "=", "True", ",", "scatter", "=", "False", ",", "linestyle", "...
Create yearly plots of a variable in a DataFrame, optionally renormalized. It is assumed that the DataFrame index is datetime.
[ "Create", "yearly", "plots", "of", "a", "variable", "in", "a", "DataFrame", "optionally", "renormalized", ".", "It", "is", "assumed", "that", "the", "DataFrame", "index", "is", "datetime", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1682-L1721
246,940
wdbm/shijian
shijian.py
add_rolling_statistics_variables
def add_rolling_statistics_variables( df = None, variable = None, window = 20, upper_factor = 2, lower_factor = 2 ): """ Add rolling statistics variables derived from a specified variable in a DataFrame. """ df[variable + "_rolling_mean"] = p...
python
def add_rolling_statistics_variables( df = None, variable = None, window = 20, upper_factor = 2, lower_factor = 2 ): """ Add rolling statistics variables derived from a specified variable in a DataFrame. """ df[variable + "_rolling_mean"] = p...
[ "def", "add_rolling_statistics_variables", "(", "df", "=", "None", ",", "variable", "=", "None", ",", "window", "=", "20", ",", "upper_factor", "=", "2", ",", "lower_factor", "=", "2", ")", ":", "df", "[", "variable", "+", "\"_rolling_mean\"", "]", "=", ...
Add rolling statistics variables derived from a specified variable in a DataFrame.
[ "Add", "rolling", "statistics", "variables", "derived", "from", "a", "specified", "variable", "in", "a", "DataFrame", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1723-L1738
246,941
wdbm/shijian
shijian.py
rescale_variables
def rescale_variables( df, variables_include = [], variables_exclude = [] ): """ Rescale variables in a DataFrame, excluding variables with NaNs and strings, excluding specified variables, and including specified variables. """ variables_not_rescale = variables_exclude variables_...
python
def rescale_variables( df, variables_include = [], variables_exclude = [] ): """ Rescale variables in a DataFrame, excluding variables with NaNs and strings, excluding specified variables, and including specified variables. """ variables_not_rescale = variables_exclude variables_...
[ "def", "rescale_variables", "(", "df", ",", "variables_include", "=", "[", "]", ",", "variables_exclude", "=", "[", "]", ")", ":", "variables_not_rescale", "=", "variables_exclude", "variables_not_rescale", ".", "extend", "(", "df", ".", "columns", "[", "df", ...
Rescale variables in a DataFrame, excluding variables with NaNs and strings, excluding specified variables, and including specified variables.
[ "Rescale", "variables", "in", "a", "DataFrame", "excluding", "variables", "with", "NaNs", "and", "strings", "excluding", "specified", "variables", "and", "including", "specified", "variables", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1740-L1756
246,942
wdbm/shijian
shijian.py
histogram_hour_counts
def histogram_hour_counts( df, variable ): """ Create a day-long histogram of counts of the variable for each hour. It is assumed that the DataFrame index is datetime and that the variable `hour` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: l...
python
def histogram_hour_counts( df, variable ): """ Create a day-long histogram of counts of the variable for each hour. It is assumed that the DataFrame index is datetime and that the variable `hour` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: l...
[ "def", "histogram_hour_counts", "(", "df", ",", "variable", ")", ":", "if", "not", "df", ".", "index", ".", "dtype", "in", "[", "\"datetime64[ns]\"", ",", "\"<M8[ns]\"", ",", "\">M8[ns]\"", "]", ":", "log", ".", "error", "(", "\"index is not datetime\"", ")"...
Create a day-long histogram of counts of the variable for each hour. It is assumed that the DataFrame index is datetime and that the variable `hour` exists.
[ "Create", "a", "day", "-", "long", "histogram", "of", "counts", "of", "the", "variable", "for", "each", "hour", ".", "It", "is", "assumed", "that", "the", "DataFrame", "index", "is", "datetime", "and", "that", "the", "variable", "hour", "exists", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1758-L1771
246,943
wdbm/shijian
shijian.py
histogram_day_counts
def histogram_day_counts( df, variable ): """ Create a week-long histogram of counts of the variable for each day. It is assumed that the DataFrame index is datetime and that the variable `weekday_name` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: ...
python
def histogram_day_counts( df, variable ): """ Create a week-long histogram of counts of the variable for each day. It is assumed that the DataFrame index is datetime and that the variable `weekday_name` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: ...
[ "def", "histogram_day_counts", "(", "df", ",", "variable", ")", ":", "if", "not", "df", ".", "index", ".", "dtype", "in", "[", "\"datetime64[ns]\"", ",", "\"<M8[ns]\"", ",", "\">M8[ns]\"", "]", ":", "log", ".", "error", "(", "\"index is not datetime\"", ")",...
Create a week-long histogram of counts of the variable for each day. It is assumed that the DataFrame index is datetime and that the variable `weekday_name` exists.
[ "Create", "a", "week", "-", "long", "histogram", "of", "counts", "of", "the", "variable", "for", "each", "day", ".", "It", "is", "assumed", "that", "the", "DataFrame", "index", "is", "datetime", "and", "that", "the", "variable", "weekday_name", "exists", "...
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1773-L1786
246,944
wdbm/shijian
shijian.py
histogram_month_counts
def histogram_month_counts( df, variable ): """ Create a year-long histogram of counts of the variable for each month. It is assumed that the DataFrame index is datetime and that the variable `month_name` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: ...
python
def histogram_month_counts( df, variable ): """ Create a year-long histogram of counts of the variable for each month. It is assumed that the DataFrame index is datetime and that the variable `month_name` exists. """ if not df.index.dtype in ["datetime64[ns]", "<M8[ns]", ">M8[ns]"]: ...
[ "def", "histogram_month_counts", "(", "df", ",", "variable", ")", ":", "if", "not", "df", ".", "index", ".", "dtype", "in", "[", "\"datetime64[ns]\"", ",", "\"<M8[ns]\"", ",", "\">M8[ns]\"", "]", ":", "log", ".", "error", "(", "\"index is not datetime\"", ")...
Create a year-long histogram of counts of the variable for each month. It is assumed that the DataFrame index is datetime and that the variable `month_name` exists.
[ "Create", "a", "year", "-", "long", "histogram", "of", "counts", "of", "the", "variable", "for", "each", "month", ".", "It", "is", "assumed", "that", "the", "DataFrame", "index", "is", "datetime", "and", "that", "the", "variable", "month_name", "exists", "...
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1788-L1801
246,945
wdbm/shijian
shijian.py
setup_Jupyter
def setup_Jupyter(): """ Set up a Jupyter notebook with a few defaults. """ sns.set(context = "paper", font = "monospace") warnings.filterwarnings("ignore") pd.set_option("display.max_rows", 500) pd.set_option("display.max_columns", 500) plt.rcParams["figure.figsize"] = (17, 10)
python
def setup_Jupyter(): """ Set up a Jupyter notebook with a few defaults. """ sns.set(context = "paper", font = "monospace") warnings.filterwarnings("ignore") pd.set_option("display.max_rows", 500) pd.set_option("display.max_columns", 500) plt.rcParams["figure.figsize"] = (17, 10)
[ "def", "setup_Jupyter", "(", ")", ":", "sns", ".", "set", "(", "context", "=", "\"paper\"", ",", "font", "=", "\"monospace\"", ")", "warnings", ".", "filterwarnings", "(", "\"ignore\"", ")", "pd", ".", "set_option", "(", "\"display.max_rows\"", ",", "500", ...
Set up a Jupyter notebook with a few defaults.
[ "Set", "up", "a", "Jupyter", "notebook", "with", "a", "few", "defaults", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L1803-L1811
246,946
wdbm/shijian
shijian.py
List_Consensus.ensure_size
def ensure_size( self, size = None ): """ This function removes the least frequent elements until the size constraint is met. """ if size is None: size = self.size_constraint while sys.getsizeof(self) > size: element_frequen...
python
def ensure_size( self, size = None ): """ This function removes the least frequent elements until the size constraint is met. """ if size is None: size = self.size_constraint while sys.getsizeof(self) > size: element_frequen...
[ "def", "ensure_size", "(", "self", ",", "size", "=", "None", ")", ":", "if", "size", "is", "None", ":", "size", "=", "self", ".", "size_constraint", "while", "sys", ".", "getsizeof", "(", "self", ")", ">", "size", ":", "element_frequencies", "=", "coll...
This function removes the least frequent elements until the size constraint is met.
[ "This", "function", "removes", "the", "least", "frequent", "elements", "until", "the", "size", "constraint", "is", "met", "." ]
ad6aea877e1eb99fe148127ea185f39f1413ed4f
https://github.com/wdbm/shijian/blob/ad6aea877e1eb99fe148127ea185f39f1413ed4f/shijian.py#L919-L932
246,947
ONSdigital/sdx-common
sdx/common/logger_config.py
logger_initial_config
def logger_initial_config(service_name=None, log_level=None, logger_format=None, logger_date_format=None): '''Set initial logging configurations. :param service_name: Name of the service :type logger: String :param log_level...
python
def logger_initial_config(service_name=None, log_level=None, logger_format=None, logger_date_format=None): '''Set initial logging configurations. :param service_name: Name of the service :type logger: String :param log_level...
[ "def", "logger_initial_config", "(", "service_name", "=", "None", ",", "log_level", "=", "None", ",", "logger_format", "=", "None", ",", "logger_date_format", "=", "None", ")", ":", "if", "not", "log_level", ":", "log_level", "=", "os", ".", "getenv", "(", ...
Set initial logging configurations. :param service_name: Name of the service :type logger: String :param log_level: A string or integer corresponding to a Python logging level :type log_level: String :param logger_format: A string defining the format of the logs :type log_level: String :...
[ "Set", "initial", "logging", "configurations", "." ]
815f6a116d41fddae182943d821dc5f582a9af69
https://github.com/ONSdigital/sdx-common/blob/815f6a116d41fddae182943d821dc5f582a9af69/sdx/common/logger_config.py#L5-L38
246,948
smartmob-project/smartmob-agent
smartmob_agent/__init__.py
access_log_middleware
async def access_log_middleware(app, handler): """Log each request in structured event log.""" event_log = app.get('smartmob.event_log') or structlog.get_logger() clock = app.get('smartmob.clock') or timeit.default_timer # Keep the request arrival time to ensure we get intuitive logging of # event...
python
async def access_log_middleware(app, handler): """Log each request in structured event log.""" event_log = app.get('smartmob.event_log') or structlog.get_logger() clock = app.get('smartmob.clock') or timeit.default_timer # Keep the request arrival time to ensure we get intuitive logging of # event...
[ "async", "def", "access_log_middleware", "(", "app", ",", "handler", ")", ":", "event_log", "=", "app", ".", "get", "(", "'smartmob.event_log'", ")", "or", "structlog", ".", "get_logger", "(", ")", "clock", "=", "app", ".", "get", "(", "'smartmob.clock'", ...
Log each request in structured event log.
[ "Log", "each", "request", "in", "structured", "event", "log", "." ]
4039f577ab7230d135f00df68c611a51e45ddbc7
https://github.com/smartmob-project/smartmob-agent/blob/4039f577ab7230d135f00df68c611a51e45ddbc7/smartmob_agent/__init__.py#L95-L139
246,949
jtpaasch/simplygithub
simplygithub/internals/merges.py
merge
def merge(profile, head, base, commit_message=None): """Merge the head of a branch into the base branch. Args: profile A profile generated from ``simplygithub.authentication.profile``. Such profiles tell this module (i) the ``repo`` to connect to, and (ii) the ``tok...
python
def merge(profile, head, base, commit_message=None): """Merge the head of a branch into the base branch. Args: profile A profile generated from ``simplygithub.authentication.profile``. Such profiles tell this module (i) the ``repo`` to connect to, and (ii) the ``tok...
[ "def", "merge", "(", "profile", ",", "head", ",", "base", ",", "commit_message", "=", "None", ")", ":", "if", "not", "commit_message", ":", "commit_message", "=", "\"Merged \"", "+", "head", "+", "\" into \"", "+", "base", "+", "\".\"", "payload", "=", "...
Merge the head of a branch into the base branch. Args: profile A profile generated from ``simplygithub.authentication.profile``. Such profiles tell this module (i) the ``repo`` to connect to, and (ii) the ``token`` to connect with. head The head to ...
[ "Merge", "the", "head", "of", "a", "branch", "into", "the", "base", "branch", "." ]
b77506275ec276ce90879bf1ea9299a79448b903
https://github.com/jtpaasch/simplygithub/blob/b77506275ec276ce90879bf1ea9299a79448b903/simplygithub/internals/merges.py#L18-L53
246,950
soasme/rio-client
rio_client/parser.py
parse_netloc
def parse_netloc(scheme, netloc): """Parse netloc string.""" auth, _netloc = netloc.split('@') sender, token = auth.split(':') if ':' in _netloc: domain, port = _netloc.split(':') port = int(port) else: domain = _netloc if scheme == 'https': port = 443 ...
python
def parse_netloc(scheme, netloc): """Parse netloc string.""" auth, _netloc = netloc.split('@') sender, token = auth.split(':') if ':' in _netloc: domain, port = _netloc.split(':') port = int(port) else: domain = _netloc if scheme == 'https': port = 443 ...
[ "def", "parse_netloc", "(", "scheme", ",", "netloc", ")", ":", "auth", ",", "_netloc", "=", "netloc", ".", "split", "(", "'@'", ")", "sender", ",", "token", "=", "auth", ".", "split", "(", "':'", ")", "if", "':'", "in", "_netloc", ":", "domain", ",...
Parse netloc string.
[ "Parse", "netloc", "string", "." ]
c6d684c6f9deea5b43f2b05bcaf40714c48b5619
https://github.com/soasme/rio-client/blob/c6d684c6f9deea5b43f2b05bcaf40714c48b5619/rio_client/parser.py#L9-L22
246,951
soasme/rio-client
rio_client/parser.py
parse_path
def parse_path(path): """Parse path string.""" version, project = path[1:].split('/') return dict(version=int(version), project=project)
python
def parse_path(path): """Parse path string.""" version, project = path[1:].split('/') return dict(version=int(version), project=project)
[ "def", "parse_path", "(", "path", ")", ":", "version", ",", "project", "=", "path", "[", "1", ":", "]", ".", "split", "(", "'/'", ")", "return", "dict", "(", "version", "=", "int", "(", "version", ")", ",", "project", "=", "project", ")" ]
Parse path string.
[ "Parse", "path", "string", "." ]
c6d684c6f9deea5b43f2b05bcaf40714c48b5619
https://github.com/soasme/rio-client/blob/c6d684c6f9deea5b43f2b05bcaf40714c48b5619/rio_client/parser.py#L24-L27
246,952
soasme/rio-client
rio_client/parser.py
parse_dsn
def parse_dsn(dsn): """Parse dsn string.""" parsed_dsn = urlparse(dsn) parsed_path = parse_path(parsed_dsn.path) return { 'scheme': parsed_dsn.scheme, 'sender': parsed_dsn.username, 'token': parsed_dsn.password, 'domain': parsed_dsn.hostname, 'port': parsed_dsn.po...
python
def parse_dsn(dsn): """Parse dsn string.""" parsed_dsn = urlparse(dsn) parsed_path = parse_path(parsed_dsn.path) return { 'scheme': parsed_dsn.scheme, 'sender': parsed_dsn.username, 'token': parsed_dsn.password, 'domain': parsed_dsn.hostname, 'port': parsed_dsn.po...
[ "def", "parse_dsn", "(", "dsn", ")", ":", "parsed_dsn", "=", "urlparse", "(", "dsn", ")", "parsed_path", "=", "parse_path", "(", "parsed_dsn", ".", "path", ")", "return", "{", "'scheme'", ":", "parsed_dsn", ".", "scheme", ",", "'sender'", ":", "parsed_dsn"...
Parse dsn string.
[ "Parse", "dsn", "string", "." ]
c6d684c6f9deea5b43f2b05bcaf40714c48b5619
https://github.com/soasme/rio-client/blob/c6d684c6f9deea5b43f2b05bcaf40714c48b5619/rio_client/parser.py#L29-L41
246,953
jldantas/libmft
libmft/api.py
MFTHeader.create_from_binary
def create_from_binary(cls, ignore_signature_check, binary_view): '''Creates a new object MFTHeader from a binary stream. The binary stream can be represented by a byte string, bytearray or a memoryview of the bytearray. Args: binary_view (memoryview of bytearray) - A binary...
python
def create_from_binary(cls, ignore_signature_check, binary_view): '''Creates a new object MFTHeader from a binary stream. The binary stream can be represented by a byte string, bytearray or a memoryview of the bytearray. Args: binary_view (memoryview of bytearray) - A binary...
[ "def", "create_from_binary", "(", "cls", ",", "ignore_signature_check", ",", "binary_view", ")", ":", "sig", ",", "fx_offset", ",", "fx_count", ",", "lsn", ",", "seq_number", ",", "hard_link_count", ",", "first_attr_offset", ",", "usage_flags", ",", "entry_len", ...
Creates a new object MFTHeader from a binary stream. The binary stream can be represented by a byte string, bytearray or a memoryview of the bytearray. Args: binary_view (memoryview of bytearray) - A binary stream with the information of the attribute Return...
[ "Creates", "a", "new", "object", "MFTHeader", "from", "a", "binary", "stream", ".", "The", "binary", "stream", "can", "be", "represented", "by", "a", "byte", "string", "bytearray", "or", "a", "memoryview", "of", "the", "bytearray", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L279-L316
246,954
jldantas/libmft
libmft/api.py
Datastream._get_dataruns
def _get_dataruns(self): '''Returns a list of dataruns, in order. ''' if self._data_runs is None: raise DataStreamError("Resident datastream don't have dataruns") if not self._data_runs_sorted: self._data_runs.sort(key=_itemgetter(0)) self._data_runs_...
python
def _get_dataruns(self): '''Returns a list of dataruns, in order. ''' if self._data_runs is None: raise DataStreamError("Resident datastream don't have dataruns") if not self._data_runs_sorted: self._data_runs.sort(key=_itemgetter(0)) self._data_runs_...
[ "def", "_get_dataruns", "(", "self", ")", ":", "if", "self", ".", "_data_runs", "is", "None", ":", "raise", "DataStreamError", "(", "\"Resident datastream don't have dataruns\"", ")", "if", "not", "self", ".", "_data_runs_sorted", ":", "self", ".", "_data_runs", ...
Returns a list of dataruns, in order.
[ "Returns", "a", "list", "of", "dataruns", "in", "order", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L389-L399
246,955
jldantas/libmft
libmft/api.py
Datastream.add_data_attribute
def add_data_attribute(self, data_attr): '''Interprets a DATA attribute and add it to the datastream.''' if data_attr.header.attr_type_id is not AttrTypes.DATA: raise DataStreamError("Invalid attribute. A Datastream deals only with DATA attributes") if data_attr.header.attr_name != s...
python
def add_data_attribute(self, data_attr): '''Interprets a DATA attribute and add it to the datastream.''' if data_attr.header.attr_type_id is not AttrTypes.DATA: raise DataStreamError("Invalid attribute. A Datastream deals only with DATA attributes") if data_attr.header.attr_name != s...
[ "def", "add_data_attribute", "(", "self", ",", "data_attr", ")", ":", "if", "data_attr", ".", "header", ".", "attr_type_id", "is", "not", "AttrTypes", ".", "DATA", ":", "raise", "DataStreamError", "(", "\"Invalid attribute. A Datastream deals only with DATA attributes\"...
Interprets a DATA attribute and add it to the datastream.
[ "Interprets", "a", "DATA", "attribute", "and", "add", "it", "to", "the", "datastream", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L405-L427
246,956
jldantas/libmft
libmft/api.py
Datastream.add_from_datastream
def add_from_datastream(self, source_ds): '''Add information from another datastream. Verifies if the datastream added is correct and copy the relevant fields if necessary.''' if source_ds.name != self.name: raise DataStreamError("Data from a different stream 'f{source_ds.name}' cann...
python
def add_from_datastream(self, source_ds): '''Add information from another datastream. Verifies if the datastream added is correct and copy the relevant fields if necessary.''' if source_ds.name != self.name: raise DataStreamError("Data from a different stream 'f{source_ds.name}' cann...
[ "def", "add_from_datastream", "(", "self", ",", "source_ds", ")", ":", "if", "source_ds", ".", "name", "!=", "self", ".", "name", ":", "raise", "DataStreamError", "(", "\"Data from a different stream 'f{source_ds.name}' cannot be add to this stream\"", ")", "if", "self"...
Add information from another datastream. Verifies if the datastream added is correct and copy the relevant fields if necessary.
[ "Add", "information", "from", "another", "datastream", ".", "Verifies", "if", "the", "datastream", "added", "is", "correct", "and", "copy", "the", "relevant", "fields", "if", "necessary", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L429-L444
246,957
jldantas/libmft
libmft/api.py
MFTEntry.create_from_binary
def create_from_binary(cls, mft_config, binary_data, entry_number): #TODO test carefully how to find the correct index entry, specially with NTFS versions < 3 '''Creates a MFTEntry from a binary stream. It correctly process the binary data extracting the MFTHeader, all the attributes and the ...
python
def create_from_binary(cls, mft_config, binary_data, entry_number): #TODO test carefully how to find the correct index entry, specially with NTFS versions < 3 '''Creates a MFTEntry from a binary stream. It correctly process the binary data extracting the MFTHeader, all the attributes and the ...
[ "def", "create_from_binary", "(", "cls", ",", "mft_config", ",", "binary_data", ",", "entry_number", ")", ":", "#TODO test carefully how to find the correct index entry, specially with NTFS versions < 3", "bin_view", "=", "memoryview", "(", "binary_data", ")", "entry", "=", ...
Creates a MFTEntry from a binary stream. It correctly process the binary data extracting the MFTHeader, all the attributes and the slack information from the binary stream. The binary data WILL be changed to apply the fixup array. Args: mft_config (:obj:`MFTConfig`) - An in...
[ "Creates", "a", "MFTEntry", "from", "a", "binary", "stream", ".", "It", "correctly", "process", "the", "binary", "data", "extracting", "the", "MFTHeader", "all", "the", "attributes", "and", "the", "slack", "information", "from", "the", "binary", "stream", "." ...
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L588-L632
246,958
jldantas/libmft
libmft/api.py
MFTEntry._find_datastream
def _find_datastream(self, name): """Find and return if a datastream exists, by name.""" for stream in self.data_streams: #search to see if this is a new datastream or a known one if stream.name == name: return stream return None
python
def _find_datastream(self, name): """Find and return if a datastream exists, by name.""" for stream in self.data_streams: #search to see if this is a new datastream or a known one if stream.name == name: return stream return None
[ "def", "_find_datastream", "(", "self", ",", "name", ")", ":", "for", "stream", "in", "self", ".", "data_streams", ":", "#search to see if this is a new datastream or a known one", "if", "stream", ".", "name", "==", "name", ":", "return", "stream", "return", "None...
Find and return if a datastream exists, by name.
[ "Find", "and", "return", "if", "a", "datastream", "exists", "by", "name", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L634-L639
246,959
jldantas/libmft
libmft/api.py
MFTEntry._add_data_attribute
def _add_data_attribute(self, data_attr): """Add a data attribute to the datastream structure. Data attributes require processing before they can be interpreted as datastream. This function grants that it is adding the attribute to the correct datastream or creating a new datastream if ...
python
def _add_data_attribute(self, data_attr): """Add a data attribute to the datastream structure. Data attributes require processing before they can be interpreted as datastream. This function grants that it is adding the attribute to the correct datastream or creating a new datastream if ...
[ "def", "_add_data_attribute", "(", "self", ",", "data_attr", ")", ":", "attr_name", "=", "data_attr", ".", "header", ".", "attr_name", "stream", "=", "self", ".", "_find_datastream", "(", "attr_name", ")", "if", "stream", "is", "None", ":", "stream", "=", ...
Add a data attribute to the datastream structure. Data attributes require processing before they can be interpreted as datastream. This function grants that it is adding the attribute to the correct datastream or creating a new datastream if necessary.
[ "Add", "a", "data", "attribute", "to", "the", "datastream", "structure", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L641-L654
246,960
jldantas/libmft
libmft/api.py
MFTEntry._load_attributes
def _load_attributes(self, mft_config, attrs_view): '''Loads all the attributes of an entry. Once executed, all the attributes should have been loaded in the attribute *attrs* instance attribute. Args: mft_config (:obj:`MFTConfig`) - An instance of MFTConfig, as this tells ...
python
def _load_attributes(self, mft_config, attrs_view): '''Loads all the attributes of an entry. Once executed, all the attributes should have been loaded in the attribute *attrs* instance attribute. Args: mft_config (:obj:`MFTConfig`) - An instance of MFTConfig, as this tells ...
[ "def", "_load_attributes", "(", "self", ",", "mft_config", ",", "attrs_view", ")", ":", "offset", "=", "0", "load_attrs", "=", "mft_config", ".", "attribute_load_list", "while", "(", "attrs_view", "[", "offset", ":", "offset", "+", "4", "]", "!=", "b'\\xff\\...
Loads all the attributes of an entry. Once executed, all the attributes should have been loaded in the attribute *attrs* instance attribute. Args: mft_config (:obj:`MFTConfig`) - An instance of MFTConfig, as this tells how the library will interpret data. ...
[ "Loads", "all", "the", "attributes", "of", "an", "entry", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L656-L681
246,961
jldantas/libmft
libmft/api.py
MFTEntry.merge_entries
def merge_entries(self, source_entry): '''Merge two entries. Allow the merging of two MFTEntries copying the attributes to the correct place and the datastreams. Args: source_entry (:obj:`MFTEntry`) - Source entry where the data will be copied from '...
python
def merge_entries(self, source_entry): '''Merge two entries. Allow the merging of two MFTEntries copying the attributes to the correct place and the datastreams. Args: source_entry (:obj:`MFTEntry`) - Source entry where the data will be copied from '...
[ "def", "merge_entries", "(", "self", ",", "source_entry", ")", ":", "#TODO should we change this to an overloaded iadd?", "#TODO I really don't like this. We are spending cycles to load things that are going to be discarted. Check another way.", "#copy the attributes", "for", "list_attr", ...
Merge two entries. Allow the merging of two MFTEntries copying the attributes to the correct place and the datastreams. Args: source_entry (:obj:`MFTEntry`) - Source entry where the data will be copied from
[ "Merge", "two", "entries", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L683-L705
246,962
jldantas/libmft
libmft/api.py
MFTEntry.get_datastream_names
def get_datastream_names(self): '''Returns a set with the datastream names. If there is no datastream, returns None ''' ads_names = set() for stream in self.data_streams: ads_names.add(stream.name) if len(ads_names): return ads_names else...
python
def get_datastream_names(self): '''Returns a set with the datastream names. If there is no datastream, returns None ''' ads_names = set() for stream in self.data_streams: ads_names.add(stream.name) if len(ads_names): return ads_names else...
[ "def", "get_datastream_names", "(", "self", ")", ":", "ads_names", "=", "set", "(", ")", "for", "stream", "in", "self", ".", "data_streams", ":", "ads_names", ".", "add", "(", "stream", ".", "name", ")", "if", "len", "(", "ads_names", ")", ":", "return...
Returns a set with the datastream names. If there is no datastream, returns None
[ "Returns", "a", "set", "with", "the", "datastream", "names", ".", "If", "there", "is", "no", "datastream", "returns", "None" ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L743-L755
246,963
jldantas/libmft
libmft/api.py
MFTEntry.get_main_filename_attr
def get_main_filename_attr(self): '''Returns the main filename attribute of the entry. As an entry can have multiple FILENAME attributes, this function allows to return the main one, i.e., the one with the lowest attribute id and the "biggest" namespace. ''' fn_attrs = s...
python
def get_main_filename_attr(self): '''Returns the main filename attribute of the entry. As an entry can have multiple FILENAME attributes, this function allows to return the main one, i.e., the one with the lowest attribute id and the "biggest" namespace. ''' fn_attrs = s...
[ "def", "get_main_filename_attr", "(", "self", ")", ":", "fn_attrs", "=", "self", ".", "get_attributes", "(", "AttrTypes", ".", "FILE_NAME", ")", "high_attr_id", "=", "0xFFFFFFFF", "main_fn", "=", "None", "if", "fn_attrs", "is", "not", "None", ":", "#search for...
Returns the main filename attribute of the entry. As an entry can have multiple FILENAME attributes, this function allows to return the main one, i.e., the one with the lowest attribute id and the "biggest" namespace.
[ "Returns", "the", "main", "filename", "attribute", "of", "the", "entry", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L757-L783
246,964
jldantas/libmft
libmft/api.py
MFT._load_relationship_info
def _load_relationship_info(self): """Maps parent and child entries in the MFT. Because the library expects the MFT file to be provided, it doesn't have access to anything non-resident. Because of that, if we want all the information related to the entry, it is necessary to visit all th...
python
def _load_relationship_info(self): """Maps parent and child entries in the MFT. Because the library expects the MFT file to be provided, it doesn't have access to anything non-resident. Because of that, if we want all the information related to the entry, it is necessary to visit all th...
[ "def", "_load_relationship_info", "(", "self", ")", ":", "mft_entry_size", "=", "self", ".", "mft_entry_size", "fp", "=", "self", ".", "file_pointer", "record_n", "=", "0", "#define the minimum amount that needs to be read", "base_struct", "=", "struct", ".", "Struct"...
Maps parent and child entries in the MFT. Because the library expects the MFT file to be provided, it doesn't have access to anything non-resident. Because of that, if we want all the information related to the entry, it is necessary to visit all the entries and map the relationship bet...
[ "Maps", "parent", "and", "child", "entries", "in", "the", "MFT", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L841-L881
246,965
jldantas/libmft
libmft/api.py
MFT._compute_full_path
def _compute_full_path(self, fn_parent_ref, fn_parent_seq): '''Based on the parent reference and sequence, computes the full path. The majority of the files in a filesystem has a very small amount of parent directories. By definition, a filesystem is expected to have much smaller amount...
python
def _compute_full_path(self, fn_parent_ref, fn_parent_seq): '''Based on the parent reference and sequence, computes the full path. The majority of the files in a filesystem has a very small amount of parent directories. By definition, a filesystem is expected to have much smaller amount...
[ "def", "_compute_full_path", "(", "self", ",", "fn_parent_ref", ",", "fn_parent_seq", ")", ":", "names", "=", "[", "]", "root_id", "=", "5", "index", ",", "seq", "=", "fn_parent_ref", ",", "fn_parent_seq", "is_orphan", "=", "False", "#search until hit the root e...
Based on the parent reference and sequence, computes the full path. The majority of the files in a filesystem has a very small amount of parent directories. By definition, a filesystem is expected to have much smaller amount of directories than files. As such we use a function with the ...
[ "Based", "on", "the", "parent", "reference", "and", "sequence", "computes", "the", "full", "path", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L903-L943
246,966
jldantas/libmft
libmft/api.py
MFT.get_full_path
def get_full_path(self, fn_attr): '''Returns the full path of a FILENAME. The NTFS filesystem allows for things called hardlinks. Hard links are saved, internally, as different filename attributes. Because of this, an entry can, when dealing with full paths, have multiple full paths. ...
python
def get_full_path(self, fn_attr): '''Returns the full path of a FILENAME. The NTFS filesystem allows for things called hardlinks. Hard links are saved, internally, as different filename attributes. Because of this, an entry can, when dealing with full paths, have multiple full paths. ...
[ "def", "get_full_path", "(", "self", ",", "fn_attr", ")", ":", "if", "fn_attr", ".", "header", ".", "attr_type_id", "is", "not", "AttrTypes", ".", "FILE_NAME", ":", "raise", "MFTError", "(", "\"Need a filename attribute to compute full path.\"", ")", "orphan", ","...
Returns the full path of a FILENAME. The NTFS filesystem allows for things called hardlinks. Hard links are saved, internally, as different filename attributes. Because of this, an entry can, when dealing with full paths, have multiple full paths. As such, this function receives a fn_a...
[ "Returns", "the", "full", "path", "of", "a", "FILENAME", "." ]
65a988605fe7663b788bd81dcb52c0a4eaad1549
https://github.com/jldantas/libmft/blob/65a988605fe7663b788bd81dcb52c0a4eaad1549/libmft/api.py#L946-L972
246,967
radjkarl/appBase
appbase/mainWindowRessources/menubar.py
MenuBar.setFullscreen
def setFullscreen(self, fullscreen): """toggle between fullscreen and normal window""" if not fullscreen: self.ckBox_fullscreen.setChecked(False) self.parent().showNormal() else: self.ckBox_fullscreen.setChecked(True) self.parent().showFullS...
python
def setFullscreen(self, fullscreen): """toggle between fullscreen and normal window""" if not fullscreen: self.ckBox_fullscreen.setChecked(False) self.parent().showNormal() else: self.ckBox_fullscreen.setChecked(True) self.parent().showFullS...
[ "def", "setFullscreen", "(", "self", ",", "fullscreen", ")", ":", "if", "not", "fullscreen", ":", "self", ".", "ckBox_fullscreen", ".", "setChecked", "(", "False", ")", "self", ".", "parent", "(", ")", ".", "showNormal", "(", ")", "else", ":", "self", ...
toggle between fullscreen and normal window
[ "toggle", "between", "fullscreen", "and", "normal", "window" ]
72b514e6dee7c083f01a2d0b2cc93d46df55bdcb
https://github.com/radjkarl/appBase/blob/72b514e6dee7c083f01a2d0b2cc93d46df55bdcb/appbase/mainWindowRessources/menubar.py#L165-L172
246,968
dhain/potpy
potpy/template.py
Template.match
def match(self, string): """Match a string against the template. If the string matches the template, return a dict mapping template parameter names to converted values, otherwise return ``None``. >>> t = Template('Hello my name is {name}!') >>> t.match('Hello my name is David!'...
python
def match(self, string): """Match a string against the template. If the string matches the template, return a dict mapping template parameter names to converted values, otherwise return ``None``. >>> t = Template('Hello my name is {name}!') >>> t.match('Hello my name is David!'...
[ "def", "match", "(", "self", ",", "string", ")", ":", "m", "=", "self", ".", "regex", ".", "match", "(", "string", ")", "if", "m", ":", "c", "=", "self", ".", "type_converters", "return", "dict", "(", "(", "k", ",", "c", "[", "k", "]", "(", "...
Match a string against the template. If the string matches the template, return a dict mapping template parameter names to converted values, otherwise return ``None``. >>> t = Template('Hello my name is {name}!') >>> t.match('Hello my name is David!') {'name': 'David'} ...
[ "Match", "a", "string", "against", "the", "template", "." ]
e39a5a84f763fbf144b07a620afb02a5ff3741c9
https://github.com/dhain/potpy/blob/e39a5a84f763fbf144b07a620afb02a5ff3741c9/potpy/template.py#L103-L119
246,969
svetlyak40wt/twiggy-goodies
twiggy_goodies/django_rq.py
job
def job(func_or_queue, connection=None, *args, **kwargs): """This decorator does all what django_rq's one, plus it group all logged messages using uuid and sets job_name field as well.""" decorated_func = _job(func_or_queue, connection=connection, *args, **kwargs) if callable(func_or_queue): ...
python
def job(func_or_queue, connection=None, *args, **kwargs): """This decorator does all what django_rq's one, plus it group all logged messages using uuid and sets job_name field as well.""" decorated_func = _job(func_or_queue, connection=connection, *args, **kwargs) if callable(func_or_queue): ...
[ "def", "job", "(", "func_or_queue", ",", "connection", "=", "None", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "decorated_func", "=", "_job", "(", "func_or_queue", ",", "connection", "=", "connection", ",", "*", "args", ",", "*", "*", "kwargs...
This decorator does all what django_rq's one, plus it group all logged messages using uuid and sets job_name field as well.
[ "This", "decorator", "does", "all", "what", "django_rq", "s", "one", "plus", "it", "group", "all", "logged", "messages", "using", "uuid", "and", "sets", "job_name", "field", "as", "well", "." ]
71528d5959fab81eb8d0e4373f20d37a013ac00e
https://github.com/svetlyak40wt/twiggy-goodies/blob/71528d5959fab81eb8d0e4373f20d37a013ac00e/twiggy_goodies/django_rq.py#L10-L32
246,970
chrisdrackett/django-support
support/templatetags/support_tags.py
active
def active(parser, token): """ tag to determine if a link is to the current page, and if it is, sets 'link_active' to True in the context. Use: {% active path view strict[True,False] arg1 arg2 arg3 %} path: the path to check. Generally this is just 'request.path' ...
python
def active(parser, token): """ tag to determine if a link is to the current page, and if it is, sets 'link_active' to True in the context. Use: {% active path view strict[True,False] arg1 arg2 arg3 %} path: the path to check. Generally this is just 'request.path' ...
[ "def", "active", "(", "parser", ",", "token", ")", ":", "args", "=", "token", ".", "split_contents", "(", ")", "path", "=", "args", "[", "1", "]", "view", "=", "args", "[", "2", "]", ".", "replace", "(", "'\"'", ",", "''", ")", ".", "replace", ...
tag to determine if a link is to the current page, and if it is, sets 'link_active' to True in the context. Use: {% active path view strict[True,False] arg1 arg2 arg3 %} path: the path to check. Generally this is just 'request.path' view: the view that you want ...
[ "tag", "to", "determine", "if", "a", "link", "is", "to", "the", "current", "page", "and", "if", "it", "is", "sets", "link_active", "to", "True", "in", "the", "context", "." ]
a4f29421a31797e0b069637a0afec85328b4f0ca
https://github.com/chrisdrackett/django-support/blob/a4f29421a31797e0b069637a0afec85328b4f0ca/support/templatetags/support_tags.py#L22-L55
246,971
chrisdrackett/django-support
support/templatetags/support_tags.py
full_width_svg
def full_width_svg(url, width, height, alt_text=None): ''' Helper to render an SVG that will size to fill its element while keeping its dimentions. ''' return { 'ratio': str((float(height)/float(width))*100)[:2], 'url': url, 'alt_text': alt_text }
python
def full_width_svg(url, width, height, alt_text=None): ''' Helper to render an SVG that will size to fill its element while keeping its dimentions. ''' return { 'ratio': str((float(height)/float(width))*100)[:2], 'url': url, 'alt_text': alt_text }
[ "def", "full_width_svg", "(", "url", ",", "width", ",", "height", ",", "alt_text", "=", "None", ")", ":", "return", "{", "'ratio'", ":", "str", "(", "(", "float", "(", "height", ")", "/", "float", "(", "width", ")", ")", "*", "100", ")", "[", ":"...
Helper to render an SVG that will size to fill its element while keeping its dimentions.
[ "Helper", "to", "render", "an", "SVG", "that", "will", "size", "to", "fill", "its", "element", "while", "keeping", "its", "dimentions", "." ]
a4f29421a31797e0b069637a0afec85328b4f0ca
https://github.com/chrisdrackett/django-support/blob/a4f29421a31797e0b069637a0afec85328b4f0ca/support/templatetags/support_tags.py#L160-L170
246,972
mirceaulinic/pypluribus
pyPluribus/device.py
PluribusDevice.open
def open(self): """Opens a SSH connection with a Pluribus machine.""" self._connection = paramiko.SSHClient() self._connection.set_missing_host_key_policy(paramiko.AutoAddPolicy()) try: self._connection.connect(hostname=self._hostname, use...
python
def open(self): """Opens a SSH connection with a Pluribus machine.""" self._connection = paramiko.SSHClient() self._connection.set_missing_host_key_policy(paramiko.AutoAddPolicy()) try: self._connection.connect(hostname=self._hostname, use...
[ "def", "open", "(", "self", ")", ":", "self", ".", "_connection", "=", "paramiko", ".", "SSHClient", "(", ")", "self", ".", "_connection", ".", "set_missing_host_key_policy", "(", "paramiko", ".", "AutoAddPolicy", "(", ")", ")", "try", ":", "self", ".", ...
Opens a SSH connection with a Pluribus machine.
[ "Opens", "a", "SSH", "connection", "with", "a", "Pluribus", "machine", "." ]
99bb9b6de40a0e465e3f0e6636b26acdeabbbd90
https://github.com/mirceaulinic/pypluribus/blob/99bb9b6de40a0e465e3f0e6636b26acdeabbbd90/pyPluribus/device.py#L57-L78
246,973
mirceaulinic/pypluribus
pyPluribus/device.py
PluribusDevice.close
def close(self): """Closes the SSH connection if the connection is UP.""" if not self.connected: return None if self.config is not None: if self.config.changed() and not self.config.committed(): try: self.config.discard() # if configur...
python
def close(self): """Closes the SSH connection if the connection is UP.""" if not self.connected: return None if self.config is not None: if self.config.changed() and not self.config.committed(): try: self.config.discard() # if configur...
[ "def", "close", "(", "self", ")", ":", "if", "not", "self", ".", "connected", ":", "return", "None", "if", "self", ".", "config", "is", "not", "None", ":", "if", "self", ".", "config", ".", "changed", "(", ")", "and", "not", "self", ".", "config", ...
Closes the SSH connection if the connection is UP.
[ "Closes", "the", "SSH", "connection", "if", "the", "connection", "is", "UP", "." ]
99bb9b6de40a0e465e3f0e6636b26acdeabbbd90
https://github.com/mirceaulinic/pypluribus/blob/99bb9b6de40a0e465e3f0e6636b26acdeabbbd90/pyPluribus/device.py#L80-L94
246,974
mirceaulinic/pypluribus
pyPluribus/device.py
PluribusDevice.cli
def cli(self, command): """ Executes a command and returns raw output from the CLI. :param command: Command to be executed on the CLI. :raise pyPluribus.exceptions.TimeoutError: when execution of the command exceeds the timeout :raise pyPluribus.exceptions.CommandExecutionError:...
python
def cli(self, command): """ Executes a command and returns raw output from the CLI. :param command: Command to be executed on the CLI. :raise pyPluribus.exceptions.TimeoutError: when execution of the command exceeds the timeout :raise pyPluribus.exceptions.CommandExecutionError:...
[ "def", "cli", "(", "self", ",", "command", ")", ":", "if", "not", "self", ".", "connected", ":", "raise", "pyPluribus", ".", "exceptions", ".", "ConnectionError", "(", "\"Not connected to the deivce.\"", ")", "cli_output", "=", "''", "ssh_session", "=", "self"...
Executes a command and returns raw output from the CLI. :param command: Command to be executed on the CLI. :raise pyPluribus.exceptions.TimeoutError: when execution of the command exceeds the timeout :raise pyPluribus.exceptions.CommandExecutionError: when not able to retrieve the output ...
[ "Executes", "a", "command", "and", "returns", "raw", "output", "from", "the", "CLI", "." ]
99bb9b6de40a0e465e3f0e6636b26acdeabbbd90
https://github.com/mirceaulinic/pypluribus/blob/99bb9b6de40a0e465e3f0e6636b26acdeabbbd90/pyPluribus/device.py#L98-L145
246,975
OpenGov/python_data_wrap
datawrap/tableloader.py
get_data_xlsx
def get_data_xlsx(file_name, file_contents=None, on_demand=False): ''' Loads the new excel format files. Old format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. ...
python
def get_data_xlsx(file_name, file_contents=None, on_demand=False): ''' Loads the new excel format files. Old format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. ...
[ "def", "get_data_xlsx", "(", "file_name", ",", "file_contents", "=", "None", ",", "on_demand", "=", "False", ")", ":", "return", "get_data_xls", "(", "file_name", ",", "file_contents", "=", "file_contents", ",", "on_demand", "=", "on_demand", ")" ]
Loads the new excel format files. Old format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. file_contents: The file-like object holding contents of file_name. ...
[ "Loads", "the", "new", "excel", "format", "files", ".", "Old", "format", "files", "will", "automatically", "get", "loaded", "as", "well", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L53-L65
246,976
OpenGov/python_data_wrap
datawrap/tableloader.py
get_data_xls
def get_data_xls(file_name, file_contents=None, on_demand=False): ''' Loads the old excel format files. New format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. ...
python
def get_data_xls(file_name, file_contents=None, on_demand=False): ''' Loads the old excel format files. New format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. ...
[ "def", "get_data_xls", "(", "file_name", ",", "file_contents", "=", "None", ",", "on_demand", "=", "False", ")", ":", "def", "tuple_to_iso_date", "(", "tuple_date", ")", ":", "'''\n Turns a gregorian (year, month, day, hour, minute, nearest_second) into a\n stan...
Loads the old excel format files. New format files will automatically get loaded as well. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. file_contents: The file-like object holding contents of file_name. ...
[ "Loads", "the", "old", "excel", "format", "files", ".", "New", "format", "files", "will", "automatically", "get", "loaded", "as", "well", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L145-L210
246,977
OpenGov/python_data_wrap
datawrap/tableloader.py
get_data_excel_xml
def get_data_excel_xml(file_name, file_contents=None, on_demand=False): ''' Loads xml excel format files. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. file_contents: The file-like object holding contents o...
python
def get_data_excel_xml(file_name, file_contents=None, on_demand=False): ''' Loads xml excel format files. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. file_contents: The file-like object holding contents o...
[ "def", "get_data_excel_xml", "(", "file_name", ",", "file_contents", "=", "None", ",", "on_demand", "=", "False", ")", ":", "# NOTE this method is inefficient and uses code that's not of the highest quality", "if", "file_contents", ":", "xml_file", "=", "BytesIO", "(", "f...
Loads xml excel format files. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. file_contents: The file-like object holding contents of file_name. If left as None, then file_name is directly loaded. ...
[ "Loads", "xml", "excel", "format", "files", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L239-L258
246,978
OpenGov/python_data_wrap
datawrap/tableloader.py
get_data_csv
def get_data_csv(file_name, encoding='utf-8', file_contents=None, on_demand=False): ''' Gets good old csv data from a file. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. encoding: Loads the file with the sp...
python
def get_data_csv(file_name, encoding='utf-8', file_contents=None, on_demand=False): ''' Gets good old csv data from a file. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. encoding: Loads the file with the sp...
[ "def", "get_data_csv", "(", "file_name", ",", "encoding", "=", "'utf-8'", ",", "file_contents", "=", "None", ",", "on_demand", "=", "False", ")", ":", "def", "yield_csv", "(", "csv_contents", ",", "csv_file", ")", ":", "try", ":", "for", "line", "in", "c...
Gets good old csv data from a file. Args: file_name: The name of the local file, or the holder for the extension type when the file_contents are supplied. encoding: Loads the file with the specified cell encoding. file_contents: The file-like object holding contents of file_name...
[ "Gets", "good", "old", "csv", "data", "from", "a", "file", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L260-L298
246,979
OpenGov/python_data_wrap
datawrap/tableloader.py
write
def write(data, file_name, worksheet_names=None): ''' Writes 2D tables to file. Args: data: 2D list of tables/worksheets. file_name: Name of the output file (determines type). worksheet_names: A list of worksheet names (optional). ''' if re.search(XML_EXT_REGEX, file_name): ...
python
def write(data, file_name, worksheet_names=None): ''' Writes 2D tables to file. Args: data: 2D list of tables/worksheets. file_name: Name of the output file (determines type). worksheet_names: A list of worksheet names (optional). ''' if re.search(XML_EXT_REGEX, file_name): ...
[ "def", "write", "(", "data", ",", "file_name", ",", "worksheet_names", "=", "None", ")", ":", "if", "re", ".", "search", "(", "XML_EXT_REGEX", ",", "file_name", ")", ":", "return", "write_xml", "(", "data", ",", "file_name", ",", "worksheet_names", "=", ...
Writes 2D tables to file. Args: data: 2D list of tables/worksheets. file_name: Name of the output file (determines type). worksheet_names: A list of worksheet names (optional).
[ "Writes", "2D", "tables", "to", "file", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L300-L318
246,980
OpenGov/python_data_wrap
datawrap/tableloader.py
write_xls
def write_xls(data, file_name, worksheet_names=None): ''' Writes out to old excel format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file. worksheet_names: A list of worksheet names (optional). ''' workbook = xlwt.Workbook() for sheet_index, ...
python
def write_xls(data, file_name, worksheet_names=None): ''' Writes out to old excel format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file. worksheet_names: A list of worksheet names (optional). ''' workbook = xlwt.Workbook() for sheet_index, ...
[ "def", "write_xls", "(", "data", ",", "file_name", ",", "worksheet_names", "=", "None", ")", ":", "workbook", "=", "xlwt", ".", "Workbook", "(", ")", "for", "sheet_index", ",", "sheet_data", "in", "enumerate", "(", "data", ")", ":", "if", "worksheet_names"...
Writes out to old excel format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file. worksheet_names: A list of worksheet names (optional).
[ "Writes", "out", "to", "old", "excel", "format", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L342-L361
246,981
OpenGov/python_data_wrap
datawrap/tableloader.py
write_csv
def write_csv(data, file_name, encoding='utf-8'): ''' Writes out to csv format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file. ''' name_extension = len(data) > 1 root, ext = os.path.splitext(file_name) for i, sheet in enumerate(data): ...
python
def write_csv(data, file_name, encoding='utf-8'): ''' Writes out to csv format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file. ''' name_extension = len(data) > 1 root, ext = os.path.splitext(file_name) for i, sheet in enumerate(data): ...
[ "def", "write_csv", "(", "data", ",", "file_name", ",", "encoding", "=", "'utf-8'", ")", ":", "name_extension", "=", "len", "(", "data", ")", ">", "1", "root", ",", "ext", "=", "os", ".", "path", ".", "splitext", "(", "file_name", ")", "for", "i", ...
Writes out to csv format. Args: data: 2D list of tables/worksheets. file_name: Name of the output file.
[ "Writes", "out", "to", "csv", "format", "." ]
7de38bb30d7a500adc336a4a7999528d753e5600
https://github.com/OpenGov/python_data_wrap/blob/7de38bb30d7a500adc336a4a7999528d753e5600/datawrap/tableloader.py#L363-L379
246,982
scdoshi/django-bits
bits/general.py
get_group
def get_group(name): """ Return group with given name, if it exists. Check cache first. """ group = cache.get('bits.general.group_%s' % name) if not group: group = Group.objects.get(name=name) cache.set('bits.general.group_%s' % name, group, 365 * 24 * 60 * 60) return group
python
def get_group(name): """ Return group with given name, if it exists. Check cache first. """ group = cache.get('bits.general.group_%s' % name) if not group: group = Group.objects.get(name=name) cache.set('bits.general.group_%s' % name, group, 365 * 24 * 60 * 60) return group
[ "def", "get_group", "(", "name", ")", ":", "group", "=", "cache", ".", "get", "(", "'bits.general.group_%s'", "%", "name", ")", "if", "not", "group", ":", "group", "=", "Group", ".", "objects", ".", "get", "(", "name", "=", "name", ")", "cache", ".",...
Return group with given name, if it exists. Check cache first.
[ "Return", "group", "with", "given", "name", "if", "it", "exists", ".", "Check", "cache", "first", "." ]
0a2f4fd9374d2a8acb8df9a7b83eebcf2782256f
https://github.com/scdoshi/django-bits/blob/0a2f4fd9374d2a8acb8df9a7b83eebcf2782256f/bits/general.py#L57-L65
246,983
abe-winter/pg13-py
pg13/sqex.py
uniqify
def uniqify(list_): "inefficient on long lists; short lists only. preserves order." a=[] for x in list_: if x not in a: a.append(x) return a
python
def uniqify(list_): "inefficient on long lists; short lists only. preserves order." a=[] for x in list_: if x not in a: a.append(x) return a
[ "def", "uniqify", "(", "list_", ")", ":", "a", "=", "[", "]", "for", "x", "in", "list_", ":", "if", "x", "not", "in", "a", ":", "a", ".", "append", "(", "x", ")", "return", "a" ]
inefficient on long lists; short lists only. preserves order.
[ "inefficient", "on", "long", "lists", ";", "short", "lists", "only", ".", "preserves", "order", "." ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L39-L44
246,984
abe-winter/pg13-py
pg13/sqex.py
eliminate_sequential_children
def eliminate_sequential_children(paths): "helper for infer_columns. removes paths that are direct children of the n-1 or n-2 path" return [p for i,p in enumerate(paths) if not ((i>0 and paths[i-1]==p[:-1]) or (i>1 and paths[i-2]==p[:-1]))]
python
def eliminate_sequential_children(paths): "helper for infer_columns. removes paths that are direct children of the n-1 or n-2 path" return [p for i,p in enumerate(paths) if not ((i>0 and paths[i-1]==p[:-1]) or (i>1 and paths[i-2]==p[:-1]))]
[ "def", "eliminate_sequential_children", "(", "paths", ")", ":", "return", "[", "p", "for", "i", ",", "p", "in", "enumerate", "(", "paths", ")", "if", "not", "(", "(", "i", ">", "0", "and", "paths", "[", "i", "-", "1", "]", "==", "p", "[", ":", ...
helper for infer_columns. removes paths that are direct children of the n-1 or n-2 path
[ "helper", "for", "infer_columns", ".", "removes", "paths", "that", "are", "direct", "children", "of", "the", "n", "-", "1", "or", "n", "-", "2", "path" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L46-L48
246,985
abe-winter/pg13-py
pg13/sqex.py
collapse_group_expr
def collapse_group_expr(groupx,cols,ret_row): "collapses columns matching the group expression. I'm sure this is buggy; look at a real DB's imp of this." for i,col in enumerate(cols.children): if col==groupx: ret_row[i]=ret_row[i][0] return ret_row
python
def collapse_group_expr(groupx,cols,ret_row): "collapses columns matching the group expression. I'm sure this is buggy; look at a real DB's imp of this." for i,col in enumerate(cols.children): if col==groupx: ret_row[i]=ret_row[i][0] return ret_row
[ "def", "collapse_group_expr", "(", "groupx", ",", "cols", ",", "ret_row", ")", ":", "for", "i", ",", "col", "in", "enumerate", "(", "cols", ".", "children", ")", ":", "if", "col", "==", "groupx", ":", "ret_row", "[", "i", "]", "=", "ret_row", "[", ...
collapses columns matching the group expression. I'm sure this is buggy; look at a real DB's imp of this.
[ "collapses", "columns", "matching", "the", "group", "expression", ".", "I", "m", "sure", "this", "is", "buggy", ";", "look", "at", "a", "real", "DB", "s", "imp", "of", "this", "." ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L209-L213
246,986
abe-winter/pg13-py
pg13/sqex.py
NameIndexer.update_aliases
def update_aliases(aliases,aonly,x): "helper for ctor. takes AliasX or string as second arg" if isinstance(x,basestring): aliases[x]=x elif isinstance(x,sqparse2.AliasX): if not isinstance(x.alias,basestring): raise TypeError('alias not string',type(x.alias)) if isinstance(x.name,sqparse2.NameX)...
python
def update_aliases(aliases,aonly,x): "helper for ctor. takes AliasX or string as second arg" if isinstance(x,basestring): aliases[x]=x elif isinstance(x,sqparse2.AliasX): if not isinstance(x.alias,basestring): raise TypeError('alias not string',type(x.alias)) if isinstance(x.name,sqparse2.NameX)...
[ "def", "update_aliases", "(", "aliases", ",", "aonly", ",", "x", ")", ":", "if", "isinstance", "(", "x", ",", "basestring", ")", ":", "aliases", "[", "x", "]", "=", "x", "elif", "isinstance", "(", "x", ",", "sqparse2", ".", "AliasX", ")", ":", "if"...
helper for ctor. takes AliasX or string as second arg
[ "helper", "for", "ctor", ".", "takes", "AliasX", "or", "string", "as", "second", "arg" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L104-L114
246,987
abe-winter/pg13-py
pg13/sqex.py
NameIndexer.resolve_aonly
def resolve_aonly(self,tables_dict,table_ctor): "circular depends on pgmock.Table. refactor." for alias,selectx in self.aonly.items(): table = table_ctor(alias,infer_columns(selectx,tables_dict),None) table.rows = run_select(selectx,tables_dict,table_ctor) self.aonly[alias] = table self.ao...
python
def resolve_aonly(self,tables_dict,table_ctor): "circular depends on pgmock.Table. refactor." for alias,selectx in self.aonly.items(): table = table_ctor(alias,infer_columns(selectx,tables_dict),None) table.rows = run_select(selectx,tables_dict,table_ctor) self.aonly[alias] = table self.ao...
[ "def", "resolve_aonly", "(", "self", ",", "tables_dict", ",", "table_ctor", ")", ":", "for", "alias", ",", "selectx", "in", "self", ".", "aonly", ".", "items", "(", ")", ":", "table", "=", "table_ctor", "(", "alias", ",", "infer_columns", "(", "selectx",...
circular depends on pgmock.Table. refactor.
[ "circular", "depends", "on", "pgmock", ".", "Table", ".", "refactor", "." ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L134-L140
246,988
abe-winter/pg13-py
pg13/sqex.py
NameIndexer.rowget
def rowget(self,tables_dict,row_list,index): "row_list in self.row_order" tmp=row_list for i in self.index_tuple(tables_dict,index,False): tmp=tmp[i] return tmp
python
def rowget(self,tables_dict,row_list,index): "row_list in self.row_order" tmp=row_list for i in self.index_tuple(tables_dict,index,False): tmp=tmp[i] return tmp
[ "def", "rowget", "(", "self", ",", "tables_dict", ",", "row_list", ",", "index", ")", ":", "tmp", "=", "row_list", "for", "i", "in", "self", ".", "index_tuple", "(", "tables_dict", ",", "index", ",", "False", ")", ":", "tmp", "=", "tmp", "[", "i", ...
row_list in self.row_order
[ "row_list", "in", "self", ".", "row_order" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L162-L166
246,989
abe-winter/pg13-py
pg13/sqex.py
Evaluator.eval_agg_call
def eval_agg_call(self, exp): "helper for eval_callx; evaluator for CallX that consume multiple rows" if not isinstance(self.c_row,list): raise TypeError('aggregate function expected a list of rows') if len(exp.args.children)!=1: raise ValueError('aggregate function expected a single value',exp.args) ar...
python
def eval_agg_call(self, exp): "helper for eval_callx; evaluator for CallX that consume multiple rows" if not isinstance(self.c_row,list): raise TypeError('aggregate function expected a list of rows') if len(exp.args.children)!=1: raise ValueError('aggregate function expected a single value',exp.args) ar...
[ "def", "eval_agg_call", "(", "self", ",", "exp", ")", ":", "if", "not", "isinstance", "(", "self", ".", "c_row", ",", "list", ")", ":", "raise", "TypeError", "(", "'aggregate function expected a list of rows'", ")", "if", "len", "(", "exp", ".", "args", "....
helper for eval_callx; evaluator for CallX that consume multiple rows
[ "helper", "for", "eval_callx", ";", "evaluator", "for", "CallX", "that", "consume", "multiple", "rows" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L261-L271
246,990
abe-winter/pg13-py
pg13/sqex.py
Evaluator.eval_nonagg_call
def eval_nonagg_call(self, exp): "helper for eval_callx; evaluator for CallX that consume a single value" # todo: get more concrete about argument counts args=self.eval(exp.args) if exp.f=='coalesce': a,b=args # todo: does coalesce take more than 2 args? return b if a is None else a elif...
python
def eval_nonagg_call(self, exp): "helper for eval_callx; evaluator for CallX that consume a single value" # todo: get more concrete about argument counts args=self.eval(exp.args) if exp.f=='coalesce': a,b=args # todo: does coalesce take more than 2 args? return b if a is None else a elif...
[ "def", "eval_nonagg_call", "(", "self", ",", "exp", ")", ":", "# todo: get more concrete about argument counts", "args", "=", "self", ".", "eval", "(", "exp", ".", "args", ")", "if", "exp", ".", "f", "==", "'coalesce'", ":", "a", ",", "b", "=", "args", "...
helper for eval_callx; evaluator for CallX that consume a single value
[ "helper", "for", "eval_callx", ";", "evaluator", "for", "CallX", "that", "consume", "a", "single", "value" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L273-L282
246,991
abe-winter/pg13-py
pg13/sqex.py
Evaluator.eval_callx
def eval_callx(self, exp): "dispatch for CallX" # below: this isn't contains(exp,consumes_row) -- it's just checking the current expression return (self.eval_agg_call if consumes_rows(exp) else self.eval_nonagg_call)(exp)
python
def eval_callx(self, exp): "dispatch for CallX" # below: this isn't contains(exp,consumes_row) -- it's just checking the current expression return (self.eval_agg_call if consumes_rows(exp) else self.eval_nonagg_call)(exp)
[ "def", "eval_callx", "(", "self", ",", "exp", ")", ":", "# below: this isn't contains(exp,consumes_row) -- it's just checking the current expression", "return", "(", "self", ".", "eval_agg_call", "if", "consumes_rows", "(", "exp", ")", "else", "self", ".", "eval_nonagg_ca...
dispatch for CallX
[ "dispatch", "for", "CallX" ]
c78806f99f35541a8756987e86edca3438aa97f5
https://github.com/abe-winter/pg13-py/blob/c78806f99f35541a8756987e86edca3438aa97f5/pg13/sqex.py#L284-L287
246,992
alfredodeza/pecan-notario
pecan_notario/decorator.py
redirect_to_handler
def redirect_to_handler(error, location): """ Cause a requset with an error to internally redirect to a URI path. This is generally for internal use, but can be called from within a Pecan controller to trigger a validation failure from *within* the controller itself, e.g.:: @expose() ...
python
def redirect_to_handler(error, location): """ Cause a requset with an error to internally redirect to a URI path. This is generally for internal use, but can be called from within a Pecan controller to trigger a validation failure from *within* the controller itself, e.g.:: @expose() ...
[ "def", "redirect_to_handler", "(", "error", ",", "location", ")", ":", "if", "callable", "(", "location", ")", ":", "location", "=", "location", "(", ")", "request", ".", "environ", "[", "'REQUEST_METHOD'", "]", "=", "'GET'", "redirect", "(", "location", "...
Cause a requset with an error to internally redirect to a URI path. This is generally for internal use, but can be called from within a Pecan controller to trigger a validation failure from *within* the controller itself, e.g.:: @expose() @validate(some_schema, '/some/handler') def...
[ "Cause", "a", "requset", "with", "an", "error", "to", "internally", "redirect", "to", "a", "URI", "path", "." ]
824e593e0ef16f276b5636315ba88f5c63b54816
https://github.com/alfredodeza/pecan-notario/blob/824e593e0ef16f276b5636315ba88f5c63b54816/pecan_notario/decorator.py#L63-L81
246,993
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._transfer_str
def _transfer_str(self, conn, tmp, name, data): ''' transfer string to remote file ''' if type(data) == dict: data = utils.jsonify(data) afd, afile = tempfile.mkstemp() afo = os.fdopen(afd, 'w') try: afo.write(data.encode('utf8')) except: ...
python
def _transfer_str(self, conn, tmp, name, data): ''' transfer string to remote file ''' if type(data) == dict: data = utils.jsonify(data) afd, afile = tempfile.mkstemp() afo = os.fdopen(afd, 'w') try: afo.write(data.encode('utf8')) except: ...
[ "def", "_transfer_str", "(", "self", ",", "conn", ",", "tmp", ",", "name", ",", "data", ")", ":", "if", "type", "(", "data", ")", "==", "dict", ":", "data", "=", "utils", ".", "jsonify", "(", "data", ")", "afd", ",", "afile", "=", "tempfile", "."...
transfer string to remote file
[ "transfer", "string", "to", "remote", "file" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L168-L188
246,994
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._execute_module
def _execute_module(self, conn, tmp, module_name, args, async_jid=None, async_module=None, async_limit=None, inject=None): ''' runs a module that has already been transferred ''' # hack to support fireball mode if module_name == 'fireball': args = "%s password=%s" % (args, ...
python
def _execute_module(self, conn, tmp, module_name, args, async_jid=None, async_module=None, async_limit=None, inject=None): ''' runs a module that has already been transferred ''' # hack to support fireball mode if module_name == 'fireball': args = "%s password=%s" % (args, ...
[ "def", "_execute_module", "(", "self", ",", "conn", ",", "tmp", ",", "module_name", ",", "args", ",", "async_jid", "=", "None", ",", "async_module", "=", "None", ",", "async_limit", "=", "None", ",", "inject", "=", "None", ")", ":", "# hack to support fire...
runs a module that has already been transferred
[ "runs", "a", "module", "that", "has", "already", "been", "transferred" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L192-L232
246,995
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._executor
def _executor(self, host): ''' handler for multiprocessing library ''' try: exec_rc = self._executor_internal(host) #if type(exec_rc) != ReturnData and type(exec_rc) != ansible.runner.return_data.ReturnData: # raise Exception("unexpected return type: %s" % type(ex...
python
def _executor(self, host): ''' handler for multiprocessing library ''' try: exec_rc = self._executor_internal(host) #if type(exec_rc) != ReturnData and type(exec_rc) != ansible.runner.return_data.ReturnData: # raise Exception("unexpected return type: %s" % type(ex...
[ "def", "_executor", "(", "self", ",", "host", ")", ":", "try", ":", "exec_rc", "=", "self", ".", "_executor_internal", "(", "host", ")", "#if type(exec_rc) != ReturnData and type(exec_rc) != ansible.runner.return_data.ReturnData:", "# raise Exception(\"unexpected return type...
handler for multiprocessing library
[ "handler", "for", "multiprocessing", "library" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L240-L258
246,996
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._executor_internal
def _executor_internal(self, host): ''' executes any module one or more times ''' host_variables = self.inventory.get_variables(host) if self.transport in [ 'paramiko', 'ssh' ]: port = host_variables.get('ansible_ssh_port', self.remote_port) if port is None: ...
python
def _executor_internal(self, host): ''' executes any module one or more times ''' host_variables = self.inventory.get_variables(host) if self.transport in [ 'paramiko', 'ssh' ]: port = host_variables.get('ansible_ssh_port', self.remote_port) if port is None: ...
[ "def", "_executor_internal", "(", "self", ",", "host", ")", ":", "host_variables", "=", "self", ".", "inventory", ".", "get_variables", "(", "host", ")", "if", "self", ".", "transport", "in", "[", "'paramiko'", ",", "'ssh'", "]", ":", "port", "=", "host_...
executes any module one or more times
[ "executes", "any", "module", "one", "or", "more", "times" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L262-L332
246,997
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._low_level_exec_command
def _low_level_exec_command(self, conn, cmd, tmp, sudoable=False, executable=None): ''' execute a command string over SSH, return the output ''' if executable is None: executable = '/bin/sh' sudo_user = self.sudo_user rc, stdin, stdout, stderr = conn.exec_command(cmd, tmp, ...
python
def _low_level_exec_command(self, conn, cmd, tmp, sudoable=False, executable=None): ''' execute a command string over SSH, return the output ''' if executable is None: executable = '/bin/sh' sudo_user = self.sudo_user rc, stdin, stdout, stderr = conn.exec_command(cmd, tmp, ...
[ "def", "_low_level_exec_command", "(", "self", ",", "conn", ",", "cmd", ",", "tmp", ",", "sudoable", "=", "False", ",", "executable", "=", "None", ")", ":", "if", "executable", "is", "None", ":", "executable", "=", "'/bin/sh'", "sudo_user", "=", "self", ...
execute a command string over SSH, return the output
[ "execute", "a", "command", "string", "over", "SSH", "return", "the", "output" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L448-L470
246,998
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._remote_md5
def _remote_md5(self, conn, tmp, path): ''' takes a remote md5sum without requiring python, and returns 0 if no file ''' test = "rc=0; [ -r \"%s\" ] || rc=2; [ -f \"%s\" ] || rc=1" % (path,path) md5s = [ "(/usr/bin/md5sum %s 2>/dev/null)" % path, # Linux "(/sbin/md5sum ...
python
def _remote_md5(self, conn, tmp, path): ''' takes a remote md5sum without requiring python, and returns 0 if no file ''' test = "rc=0; [ -r \"%s\" ] || rc=2; [ -f \"%s\" ] || rc=1" % (path,path) md5s = [ "(/usr/bin/md5sum %s 2>/dev/null)" % path, # Linux "(/sbin/md5sum ...
[ "def", "_remote_md5", "(", "self", ",", "conn", ",", "tmp", ",", "path", ")", ":", "test", "=", "\"rc=0; [ -r \\\"%s\\\" ] || rc=2; [ -f \\\"%s\\\" ] || rc=1\"", "%", "(", "path", ",", "path", ")", "md5s", "=", "[", "\"(/usr/bin/md5sum %s 2>/dev/null)\"", "%", "pa...
takes a remote md5sum without requiring python, and returns 0 if no file
[ "takes", "a", "remote", "md5sum", "without", "requiring", "python", "and", "returns", "0", "if", "no", "file" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L474-L500
246,999
cirruscluster/cirruscluster
cirruscluster/ext/ansible/runner/__init__.py
Runner._make_tmp_path
def _make_tmp_path(self, conn): ''' make and return a temporary path on a remote box ''' basefile = 'ansible-%s-%s' % (time.time(), random.randint(0, 2**48)) basetmp = os.path.join(C.DEFAULT_REMOTE_TMP, basefile) if self.sudo and self.sudo_user != 'root': basetmp = os.path.j...
python
def _make_tmp_path(self, conn): ''' make and return a temporary path on a remote box ''' basefile = 'ansible-%s-%s' % (time.time(), random.randint(0, 2**48)) basetmp = os.path.join(C.DEFAULT_REMOTE_TMP, basefile) if self.sudo and self.sudo_user != 'root': basetmp = os.path.j...
[ "def", "_make_tmp_path", "(", "self", ",", "conn", ")", ":", "basefile", "=", "'ansible-%s-%s'", "%", "(", "time", ".", "time", "(", ")", ",", "random", ".", "randint", "(", "0", ",", "2", "**", "48", ")", ")", "basetmp", "=", "os", ".", "path", ...
make and return a temporary path on a remote box
[ "make", "and", "return", "a", "temporary", "path", "on", "a", "remote", "box" ]
977409929dd81322d886425cdced10608117d5d7
https://github.com/cirruscluster/cirruscluster/blob/977409929dd81322d886425cdced10608117d5d7/cirruscluster/ext/ansible/runner/__init__.py#L504-L519