repository_name
stringlengths
7
55
func_path_in_repository
stringlengths
4
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
75
104k
language
stringclasses
1 value
func_code_string
stringlengths
75
104k
func_code_tokens
listlengths
19
28.4k
func_documentation_string
stringlengths
1
46.9k
func_documentation_tokens
listlengths
1
1.97k
split_name
stringclasses
1 value
func_code_url
stringlengths
87
315
sashs/filebytes
filebytes/pe.py
PE._getSectionForDataDirectoryEntry
def _getSectionForDataDirectoryEntry(self, data_directory_entry, sections): """Returns the section which contains the data of DataDirectory""" for section in sections: if data_directory_entry.VirtualAddress >= section.header.VirtualAddress and \ data_directory_entry.VirtualAddres...
python
def _getSectionForDataDirectoryEntry(self, data_directory_entry, sections): """Returns the section which contains the data of DataDirectory""" for section in sections: if data_directory_entry.VirtualAddress >= section.header.VirtualAddress and \ data_directory_entry.VirtualAddres...
[ "def", "_getSectionForDataDirectoryEntry", "(", "self", ",", "data_directory_entry", ",", "sections", ")", ":", "for", "section", "in", "sections", ":", "if", "data_directory_entry", ".", "VirtualAddress", ">=", "section", ".", "header", ".", "VirtualAddress", "and"...
Returns the section which contains the data of DataDirectory
[ "Returns", "the", "section", "which", "contains", "the", "data", "of", "DataDirectory" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L541-L547
sashs/filebytes
filebytes/pe.py
PE._parseDataDirectory
def _parseDataDirectory(self, data, sections, imageNtHeaders): """Parses the entries of the DataDirectory and returns a list of the content""" data_directory_data_list = [None for i in range(15)] # parse DataDirectory[Export] export_data_directory = imageNtHeaders.header.OptionalHeader....
python
def _parseDataDirectory(self, data, sections, imageNtHeaders): """Parses the entries of the DataDirectory and returns a list of the content""" data_directory_data_list = [None for i in range(15)] # parse DataDirectory[Export] export_data_directory = imageNtHeaders.header.OptionalHeader....
[ "def", "_parseDataDirectory", "(", "self", ",", "data", ",", "sections", ",", "imageNtHeaders", ")", ":", "data_directory_data_list", "=", "[", "None", "for", "i", "in", "range", "(", "15", ")", "]", "# parse DataDirectory[Export]", "export_data_directory", "=", ...
Parses the entries of the DataDirectory and returns a list of the content
[ "Parses", "the", "entries", "of", "the", "DataDirectory", "and", "returns", "a", "list", "of", "the", "content" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L549-L571
sashs/filebytes
filebytes/pe.py
PE._parseDataDirectoryExport
def _parseDataDirectoryExport(self, data, dataDirectoryEntry, exportSection): """Parses the EmportDataDirectory and returns an instance of ExportDirectoryData""" if not exportSection: return functions = [] export_directory = IMAGE_EXPORT_DIRECTORY.from_buffer(exportSection.ra...
python
def _parseDataDirectoryExport(self, data, dataDirectoryEntry, exportSection): """Parses the EmportDataDirectory and returns an instance of ExportDirectoryData""" if not exportSection: return functions = [] export_directory = IMAGE_EXPORT_DIRECTORY.from_buffer(exportSection.ra...
[ "def", "_parseDataDirectoryExport", "(", "self", ",", "data", ",", "dataDirectoryEntry", ",", "exportSection", ")", ":", "if", "not", "exportSection", ":", "return", "functions", "=", "[", "]", "export_directory", "=", "IMAGE_EXPORT_DIRECTORY", ".", "from_buffer", ...
Parses the EmportDataDirectory and returns an instance of ExportDirectoryData
[ "Parses", "the", "EmportDataDirectory", "and", "returns", "an", "instance", "of", "ExportDirectoryData" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L573-L601
sashs/filebytes
filebytes/pe.py
PE._parseDataDirectoryImport
def _parseDataDirectoryImport(self, dataDirectoryEntry, importSection): """Parses the ImportDataDirectory and returns a list of ImportDescriptorData""" if not importSection: return raw_bytes = (c_ubyte * dataDirectoryEntry.Size).from_buffer(importSection.raw, to_offset(dataDirector...
python
def _parseDataDirectoryImport(self, dataDirectoryEntry, importSection): """Parses the ImportDataDirectory and returns a list of ImportDescriptorData""" if not importSection: return raw_bytes = (c_ubyte * dataDirectoryEntry.Size).from_buffer(importSection.raw, to_offset(dataDirector...
[ "def", "_parseDataDirectoryImport", "(", "self", ",", "dataDirectoryEntry", ",", "importSection", ")", ":", "if", "not", "importSection", ":", "return", "raw_bytes", "=", "(", "c_ubyte", "*", "dataDirectoryEntry", ".", "Size", ")", ".", "from_buffer", "(", "impo...
Parses the ImportDataDirectory and returns a list of ImportDescriptorData
[ "Parses", "the", "ImportDataDirectory", "and", "returns", "a", "list", "of", "ImportDescriptorData" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L603-L629
sashs/filebytes
filebytes/pe.py
PE.__parseThunks
def __parseThunks(self, thunkRVA, importSection): """Parses the thunks and returns a list""" offset = to_offset(thunkRVA, importSection) table_offset = 0 thunks = [] while True: thunk = IMAGE_THUNK_DATA.from_buffer(importSection.raw, offset) offset += size...
python
def __parseThunks(self, thunkRVA, importSection): """Parses the thunks and returns a list""" offset = to_offset(thunkRVA, importSection) table_offset = 0 thunks = [] while True: thunk = IMAGE_THUNK_DATA.from_buffer(importSection.raw, offset) offset += size...
[ "def", "__parseThunks", "(", "self", ",", "thunkRVA", ",", "importSection", ")", ":", "offset", "=", "to_offset", "(", "thunkRVA", ",", "importSection", ")", "table_offset", "=", "0", "thunks", "=", "[", "]", "while", "True", ":", "thunk", "=", "IMAGE_THUN...
Parses the thunks and returns a list
[ "Parses", "the", "thunks", "and", "returns", "a", "list" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L672-L687
sashs/filebytes
filebytes/pe.py
PE.__parseThunkData
def __parseThunkData(self, thunk,importSection): """Parses the data of a thunk and sets the data""" offset = to_offset(thunk.header.AddressOfData, importSection) if 0xf0000000 & thunk.header.AddressOfData == 0x80000000: thunk.ordinal = thunk.header.AddressOfData & 0x0fffffff ...
python
def __parseThunkData(self, thunk,importSection): """Parses the data of a thunk and sets the data""" offset = to_offset(thunk.header.AddressOfData, importSection) if 0xf0000000 & thunk.header.AddressOfData == 0x80000000: thunk.ordinal = thunk.header.AddressOfData & 0x0fffffff ...
[ "def", "__parseThunkData", "(", "self", ",", "thunk", ",", "importSection", ")", ":", "offset", "=", "to_offset", "(", "thunk", ".", "header", ".", "AddressOfData", ",", "importSection", ")", "if", "0xf0000000", "&", "thunk", ".", "header", ".", "AddressOfDa...
Parses the data of a thunk and sets the data
[ "Parses", "the", "data", "of", "a", "thunk", "and", "sets", "the", "data" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L689-L699
sashs/filebytes
filebytes/ctypes_helper.py
get_ptr
def get_ptr(data, offset=None, ptr_type=ctypes.c_void_p): """Returns a void pointer to the data""" ptr = ctypes.cast(ctypes.pointer(data), ctypes.c_void_p) if offset: ptr = ctypes.c_void_p(ptr.value + offset) if ptr_type != ctypes.c_void_p: ptr = ctypes.cast(ptr, ptr_type) return ...
python
def get_ptr(data, offset=None, ptr_type=ctypes.c_void_p): """Returns a void pointer to the data""" ptr = ctypes.cast(ctypes.pointer(data), ctypes.c_void_p) if offset: ptr = ctypes.c_void_p(ptr.value + offset) if ptr_type != ctypes.c_void_p: ptr = ctypes.cast(ptr, ptr_type) return ...
[ "def", "get_ptr", "(", "data", ",", "offset", "=", "None", ",", "ptr_type", "=", "ctypes", ".", "c_void_p", ")", ":", "ptr", "=", "ctypes", ".", "cast", "(", "ctypes", ".", "pointer", "(", "data", ")", ",", "ctypes", ".", "c_void_p", ")", "if", "of...
Returns a void pointer to the data
[ "Returns", "a", "void", "pointer", "to", "the", "data" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/ctypes_helper.py#L33-L43
sashs/filebytes
filebytes/ctypes_helper.py
to_ubyte_array
def to_ubyte_array(barray): """Returns a c_ubyte_array filled with the given data of a bytearray or bytes""" bs = (ctypes.c_ubyte * len(barray))() pack_into('%ds' % len(barray), bs, 0, barray) return bs
python
def to_ubyte_array(barray): """Returns a c_ubyte_array filled with the given data of a bytearray or bytes""" bs = (ctypes.c_ubyte * len(barray))() pack_into('%ds' % len(barray), bs, 0, barray) return bs
[ "def", "to_ubyte_array", "(", "barray", ")", ":", "bs", "=", "(", "ctypes", ".", "c_ubyte", "*", "len", "(", "barray", ")", ")", "(", ")", "pack_into", "(", "'%ds'", "%", "len", "(", "barray", ")", ",", "bs", ",", "0", ",", "barray", ")", "return...
Returns a c_ubyte_array filled with the given data of a bytearray or bytes
[ "Returns", "a", "c_ubyte_array", "filled", "with", "the", "given", "data", "of", "a", "bytearray", "or", "bytes" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/ctypes_helper.py#L48-L53
sashs/filebytes
filebytes/binary.py
Binary._readFile
def _readFile(self, fileName): """ Returns the bytes of the file. """ with open(fileName, 'rb') as binFile: b = binFile.read() return to_ubyte_array(b)
python
def _readFile(self, fileName): """ Returns the bytes of the file. """ with open(fileName, 'rb') as binFile: b = binFile.read() return to_ubyte_array(b)
[ "def", "_readFile", "(", "self", ",", "fileName", ")", ":", "with", "open", "(", "fileName", ",", "'rb'", ")", "as", "binFile", ":", "b", "=", "binFile", ".", "read", "(", ")", "return", "to_ubyte_array", "(", "b", ")" ]
Returns the bytes of the file.
[ "Returns", "the", "bytes", "of", "the", "file", "." ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/binary.py#L71-L77
sashs/filebytes
filebytes/elf.py
ELF._getSuitableClasses
def _getSuitableClasses(self, data): """Returns the class which holds the suitable classes for the loaded file""" classes = None if data[EI.CLASS] == ELFCLASS.BITS_32: if data[EI.DATA] == ELFDATA.LSB: classes = LSB_32 elif data[EI.DATA] == ELFDATA.MSB: ...
python
def _getSuitableClasses(self, data): """Returns the class which holds the suitable classes for the loaded file""" classes = None if data[EI.CLASS] == ELFCLASS.BITS_32: if data[EI.DATA] == ELFDATA.LSB: classes = LSB_32 elif data[EI.DATA] == ELFDATA.MSB: ...
[ "def", "_getSuitableClasses", "(", "self", ",", "data", ")", ":", "classes", "=", "None", "if", "data", "[", "EI", ".", "CLASS", "]", "==", "ELFCLASS", ".", "BITS_32", ":", "if", "data", "[", "EI", ".", "DATA", "]", "==", "ELFDATA", ".", "LSB", ":"...
Returns the class which holds the suitable classes for the loaded file
[ "Returns", "the", "class", "which", "holds", "the", "suitable", "classes", "for", "the", "loaded", "file" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L881-L896
sashs/filebytes
filebytes/elf.py
ELF._parseElfHeader
def _parseElfHeader(self, data): """Returns the elf header""" ehdr = self.__classes.EHDR.from_buffer(data) return EhdrData(header=ehdr)
python
def _parseElfHeader(self, data): """Returns the elf header""" ehdr = self.__classes.EHDR.from_buffer(data) return EhdrData(header=ehdr)
[ "def", "_parseElfHeader", "(", "self", ",", "data", ")", ":", "ehdr", "=", "self", ".", "__classes", ".", "EHDR", ".", "from_buffer", "(", "data", ")", "return", "EhdrData", "(", "header", "=", "ehdr", ")" ]
Returns the elf header
[ "Returns", "the", "elf", "header" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L898-L901
sashs/filebytes
filebytes/elf.py
ELF._parseSegments
def _parseSegments(self, data, elfHeader): """Return a list of segments""" offset = elfHeader.header.e_phoff segments = [] for i in range(elfHeader.header.e_phnum): phdr = self.__classes.PHDR.from_buffer(data, offset) segment_bytes = (c_ubyte * phdr.p_filesz).from...
python
def _parseSegments(self, data, elfHeader): """Return a list of segments""" offset = elfHeader.header.e_phoff segments = [] for i in range(elfHeader.header.e_phnum): phdr = self.__classes.PHDR.from_buffer(data, offset) segment_bytes = (c_ubyte * phdr.p_filesz).from...
[ "def", "_parseSegments", "(", "self", ",", "data", ",", "elfHeader", ")", ":", "offset", "=", "elfHeader", ".", "header", ".", "e_phoff", "segments", "=", "[", "]", "for", "i", "in", "range", "(", "elfHeader", ".", "header", ".", "e_phnum", ")", ":", ...
Return a list of segments
[ "Return", "a", "list", "of", "segments" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L903-L916
sashs/filebytes
filebytes/elf.py
ELF._parseSections
def _parseSections(self, data, elfHeader): """Returns a list of sections""" offset = elfHeader.header.e_shoff shdrs = [] for i in range(elfHeader.header.e_shnum): shdr = self.__classes.SHDR.from_buffer(data, offset) section_bytes = None ba_section_byte...
python
def _parseSections(self, data, elfHeader): """Returns a list of sections""" offset = elfHeader.header.e_shoff shdrs = [] for i in range(elfHeader.header.e_shnum): shdr = self.__classes.SHDR.from_buffer(data, offset) section_bytes = None ba_section_byte...
[ "def", "_parseSections", "(", "self", ",", "data", ",", "elfHeader", ")", ":", "offset", "=", "elfHeader", ".", "header", ".", "e_shoff", "shdrs", "=", "[", "]", "for", "i", "in", "range", "(", "elfHeader", ".", "header", ".", "e_shnum", ")", ":", "s...
Returns a list of sections
[ "Returns", "a", "list", "of", "sections" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L918-L939
sashs/filebytes
filebytes/elf.py
ELF._parseSymbols
def _parseSymbols(self, sections): """Sets a list of symbols in each DYNSYM and SYMTAB section""" for section in sections: strtab = sections[section.header.sh_link] if section.header.sh_type in (int(SHT.DYNSYM), int(SHT.SYMTAB)): section.symbols = self.__parseSymb...
python
def _parseSymbols(self, sections): """Sets a list of symbols in each DYNSYM and SYMTAB section""" for section in sections: strtab = sections[section.header.sh_link] if section.header.sh_type in (int(SHT.DYNSYM), int(SHT.SYMTAB)): section.symbols = self.__parseSymb...
[ "def", "_parseSymbols", "(", "self", ",", "sections", ")", ":", "for", "section", "in", "sections", ":", "strtab", "=", "sections", "[", "section", ".", "header", ".", "sh_link", "]", "if", "section", ".", "header", ".", "sh_type", "in", "(", "int", "(...
Sets a list of symbols in each DYNSYM and SYMTAB section
[ "Sets", "a", "list", "of", "symbols", "in", "each", "DYNSYM", "and", "SYMTAB", "section" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L941-L946
sashs/filebytes
filebytes/elf.py
ELF._parseRelocations
def _parseRelocations(self, sections): """Parses the relocations and add those to the section""" for section in sections: if section.header.sh_link != SHN.UNDEF and section.header.sh_type in (SHT.REL, SHT.RELA): symbols = sections[section.header.sh_link].symbols ...
python
def _parseRelocations(self, sections): """Parses the relocations and add those to the section""" for section in sections: if section.header.sh_link != SHN.UNDEF and section.header.sh_type in (SHT.REL, SHT.RELA): symbols = sections[section.header.sh_link].symbols ...
[ "def", "_parseRelocations", "(", "self", ",", "sections", ")", ":", "for", "section", "in", "sections", ":", "if", "section", ".", "header", ".", "sh_link", "!=", "SHN", ".", "UNDEF", "and", "section", ".", "header", ".", "sh_type", "in", "(", "SHT", "...
Parses the relocations and add those to the section
[ "Parses", "the", "relocations", "and", "add", "those", "to", "the", "section" ]
train
https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L965-L971
pyqg/pyqg
pyqg/model.py
run_with_snapshots
def run_with_snapshots(self, tsnapstart=0., tsnapint=432000.): """Run the model forward, yielding to user code at specified intervals. Parameters ---------- tsnapstart : int The timestep at which to begin yielding. tstapint : int The interval at which to...
python
def run_with_snapshots(self, tsnapstart=0., tsnapint=432000.): """Run the model forward, yielding to user code at specified intervals. Parameters ---------- tsnapstart : int The timestep at which to begin yielding. tstapint : int The interval at which to...
[ "def", "run_with_snapshots", "(", "self", ",", "tsnapstart", "=", "0.", ",", "tsnapint", "=", "432000.", ")", ":", "tsnapints", "=", "np", ".", "ceil", "(", "tsnapint", "/", "self", ".", "dt", ")", "while", "(", "self", ".", "t", "<", "self", ".", ...
Run the model forward, yielding to user code at specified intervals. Parameters ---------- tsnapstart : int The timestep at which to begin yielding. tstapint : int The interval at which to yield.
[ "Run", "the", "model", "forward", "yielding", "to", "user", "code", "at", "specified", "intervals", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L210-L228
pyqg/pyqg
pyqg/model.py
vertical_modes
def vertical_modes(self): """ Calculate standard vertical modes. Simply the eigenvectors of the stretching matrix S """ evals,evecs = np.linalg.eig(-self.S) asort = evals.argsort() # deformation wavenumbers and radii self.kdi2 = evals[asort] self.radii = np...
python
def vertical_modes(self): """ Calculate standard vertical modes. Simply the eigenvectors of the stretching matrix S """ evals,evecs = np.linalg.eig(-self.S) asort = evals.argsort() # deformation wavenumbers and radii self.kdi2 = evals[asort] self.radii = np...
[ "def", "vertical_modes", "(", "self", ")", ":", "evals", ",", "evecs", "=", "np", ".", "linalg", ".", "eig", "(", "-", "self", ".", "S", ")", "asort", "=", "evals", ".", "argsort", "(", ")", "# deformation wavenumbers and radii", "self", ".", "kdi2", "...
Calculate standard vertical modes. Simply the eigenvectors of the stretching matrix S
[ "Calculate", "standard", "vertical", "modes", ".", "Simply", "the", "eigenvectors", "of", "the", "stretching", "matrix", "S" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L236-L255
pyqg/pyqg
pyqg/model.py
modal_projection
def modal_projection(self,p,forward=True): """ Performs a field p into modal amplitudes pn using the basis [pmodes]. The inverse transform calculates p from pn""" if forward: pt = np.linalg.solve(self.pmodes[np.newaxis,np.newaxis],p.T).T else: ...
python
def modal_projection(self,p,forward=True): """ Performs a field p into modal amplitudes pn using the basis [pmodes]. The inverse transform calculates p from pn""" if forward: pt = np.linalg.solve(self.pmodes[np.newaxis,np.newaxis],p.T).T else: ...
[ "def", "modal_projection", "(", "self", ",", "p", ",", "forward", "=", "True", ")", ":", "if", "forward", ":", "pt", "=", "np", ".", "linalg", ".", "solve", "(", "self", ".", "pmodes", "[", "np", ".", "newaxis", ",", "np", ".", "newaxis", "]", ",...
Performs a field p into modal amplitudes pn using the basis [pmodes]. The inverse transform calculates p from pn
[ "Performs", "a", "field", "p", "into", "modal", "amplitudes", "pn", "using", "the", "basis", "[", "pmodes", "]", ".", "The", "inverse", "transform", "calculates", "p", "from", "pn" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L257-L267
pyqg/pyqg
pyqg/sqg_model.py
SQGModel._initialize_background
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" # background vel. if len(np.shape(self.U)) == 0: self.U = (self.U * np.ones((self.ny))) print(np.shape(self.U)) self.set_U(self.U) # the meridional PV gradients in e...
python
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" # background vel. if len(np.shape(self.U)) == 0: self.U = (self.U * np.ones((self.ny))) print(np.shape(self.U)) self.set_U(self.U) # the meridional PV gradients in e...
[ "def", "_initialize_background", "(", "self", ")", ":", "# background vel.", "if", "len", "(", "np", ".", "shape", "(", "self", ".", "U", ")", ")", "==", "0", ":", "self", ".", "U", "=", "(", "self", ".", "U", "*", "np", ".", "ones", "(", "(", ...
Set up background state (zonal flow and PV gradients).
[ "Set", "up", "background", "state", "(", "zonal", "flow", "and", "PV", "gradients", ")", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L51-L67
pyqg/pyqg
pyqg/sqg_model.py
SQGModel._initialize_inversion_matrix
def _initialize_inversion_matrix(self): """ the inversion """ # The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph self.a = np.asarray(self.Nb*np.sqrt(self.wv2i))[np.newaxis, np.newaxis, :, :]
python
def _initialize_inversion_matrix(self): """ the inversion """ # The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph self.a = np.asarray(self.Nb*np.sqrt(self.wv2i))[np.newaxis, np.newaxis, :, :]
[ "def", "_initialize_inversion_matrix", "(", "self", ")", ":", "# The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph", "self", ".", "a", "=", "np", ".", "asarray", "(", "self", ".", "Nb", "*", "np", ".", "sqrt", "(", "self", ".", "wv2i", ")", ")...
the inversion
[ "the", "inversion" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L69-L72
pyqg/pyqg
pyqg/sqg_model.py
SQGModel.set_U
def set_U(self, U): """Set background zonal flow""" self.Ubg = np.asarray(U)[np.newaxis,...]
python
def set_U(self, U): """Set background zonal flow""" self.Ubg = np.asarray(U)[np.newaxis,...]
[ "def", "set_U", "(", "self", ",", "U", ")", ":", "self", ".", "Ubg", "=", "np", ".", "asarray", "(", "U", ")", "[", "np", ".", "newaxis", ",", "...", "]" ]
Set background zonal flow
[ "Set", "background", "zonal", "flow" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L77-L79
pyqg/pyqg
pyqg/particles.py
LagrangianParticleArray2D.step_forward_with_function
def step_forward_with_function(self, uv0fun, uv1fun, dt): """Advance particles using a function to determine u and v. Parameters ---------- uv0fun : function Called like ``uv0fun(x,y)``. Should return the velocity field u, v at time t. uv1fun(x,y)...
python
def step_forward_with_function(self, uv0fun, uv1fun, dt): """Advance particles using a function to determine u and v. Parameters ---------- uv0fun : function Called like ``uv0fun(x,y)``. Should return the velocity field u, v at time t. uv1fun(x,y)...
[ "def", "step_forward_with_function", "(", "self", ",", "uv0fun", ",", "uv1fun", ",", "dt", ")", ":", "dx", ",", "dy", "=", "self", ".", "_rk4_integrate", "(", "self", ".", "x", ",", "self", ".", "y", ",", "uv0fun", ",", "uv1fun", ",", "dt", ")", "s...
Advance particles using a function to determine u and v. Parameters ---------- uv0fun : function Called like ``uv0fun(x,y)``. Should return the velocity field u, v at time t. uv1fun(x,y) : function Called like ``uv1fun(x,y)``. Should return th...
[ "Advance", "particles", "using", "a", "function", "to", "determine", "u", "and", "v", ".", "Parameters", "----------", "uv0fun", ":", "function", "Called", "like", "uv0fun", "(", "x", "y", ")", ".", "Should", "return", "the", "velocity", "field", "u", "v",...
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L65-L81
pyqg/pyqg
pyqg/particles.py
LagrangianParticleArray2D._rk4_integrate
def _rk4_integrate(self, x, y, uv0fun, uv1fun, dt): """Integrates positions x, y using velocity functions uv0fun, uv1fun. Returns dx and dy, the displacements.""" u0, v0 = uv0fun(x, y) k1u = dt*u0 k1v = dt*v0 x11 = self._wrap_x(x + 0.5*k1u) y11 = self._wrap_y(y...
python
def _rk4_integrate(self, x, y, uv0fun, uv1fun, dt): """Integrates positions x, y using velocity functions uv0fun, uv1fun. Returns dx and dy, the displacements.""" u0, v0 = uv0fun(x, y) k1u = dt*u0 k1v = dt*v0 x11 = self._wrap_x(x + 0.5*k1u) y11 = self._wrap_y(y...
[ "def", "_rk4_integrate", "(", "self", ",", "x", ",", "y", ",", "uv0fun", ",", "uv1fun", ",", "dt", ")", ":", "u0", ",", "v0", "=", "uv0fun", "(", "x", ",", "y", ")", "k1u", "=", "dt", "*", "u0", "k1v", "=", "dt", "*", "v0", "x11", "=", "sel...
Integrates positions x, y using velocity functions uv0fun, uv1fun. Returns dx and dy, the displacements.
[ "Integrates", "positions", "x", "y", "using", "velocity", "functions", "uv0fun", "uv1fun", ".", "Returns", "dx", "and", "dy", "the", "displacements", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L83-L108
pyqg/pyqg
pyqg/particles.py
LagrangianParticleArray2D._distance
def _distance(self, x0, y0, x1, y1): """Utitlity function to compute distance between points.""" dx = x1-x0 dy = y1-y0 # roll displacements across the borders if self.pix: dx[ dx > self.Lx/2 ] -= self.Lx dx[ dx < -self.Lx/2 ] += self.Lx if self.piy...
python
def _distance(self, x0, y0, x1, y1): """Utitlity function to compute distance between points.""" dx = x1-x0 dy = y1-y0 # roll displacements across the borders if self.pix: dx[ dx > self.Lx/2 ] -= self.Lx dx[ dx < -self.Lx/2 ] += self.Lx if self.piy...
[ "def", "_distance", "(", "self", ",", "x0", ",", "y0", ",", "x1", ",", "y1", ")", ":", "dx", "=", "x1", "-", "x0", "dy", "=", "y1", "-", "y0", "# roll displacements across the borders", "if", "self", ".", "pix", ":", "dx", "[", "dx", ">", "self", ...
Utitlity function to compute distance between points.
[ "Utitlity", "function", "to", "compute", "distance", "between", "points", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L124-L135
pyqg/pyqg
pyqg/particles.py
GriddedLagrangianParticleArray2D.interpolate_gridded_scalar
def interpolate_gridded_scalar(self, x, y, c, order=1, pad=1, offset=0): """Interpolate gridded scalar C to points x,y. Parameters ---------- x, y : array-like Points at which to interpolate c : array-like The scalar, assumed to be defined on the ...
python
def interpolate_gridded_scalar(self, x, y, c, order=1, pad=1, offset=0): """Interpolate gridded scalar C to points x,y. Parameters ---------- x, y : array-like Points at which to interpolate c : array-like The scalar, assumed to be defined on the ...
[ "def", "interpolate_gridded_scalar", "(", "self", ",", "x", ",", "y", ",", "c", ",", "order", "=", "1", ",", "pad", "=", "1", ",", "offset", "=", "0", ")", ":", "## no longer necessary because we accept pre-padded arrays", "# assert c.shape == (self.Ny, self.Nx), 'S...
Interpolate gridded scalar C to points x,y. Parameters ---------- x, y : array-like Points at which to interpolate c : array-like The scalar, assumed to be defined on the grid. order : int Order of interpolation pad : int ...
[ "Interpolate", "gridded", "scalar", "C", "to", "points", "x", "y", ".", "Parameters", "----------", "x", "y", ":", "array", "-", "like", "Points", "at", "which", "to", "interpolate", "c", ":", "array", "-", "like", "The", "scalar", "assumed", "to", "be",...
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L180-L218
pyqg/pyqg
pyqg/particles.py
GriddedLagrangianParticleArray2D.step_forward_with_gridded_uv
def step_forward_with_gridded_uv(self, U0, V0, U1, V1, dt, order=1): """Advance particles using a gridded velocity field. Because of the Runga-Kutta timestepping, we need two velocity fields at different times. Parameters ---------- U0, V0 : array-like ...
python
def step_forward_with_gridded_uv(self, U0, V0, U1, V1, dt, order=1): """Advance particles using a gridded velocity field. Because of the Runga-Kutta timestepping, we need two velocity fields at different times. Parameters ---------- U0, V0 : array-like ...
[ "def", "step_forward_with_gridded_uv", "(", "self", ",", "U0", ",", "V0", ",", "U1", ",", "V1", ",", "dt", ",", "order", "=", "1", ")", ":", "# create interpolation functions which return u and v", "# pre-pad arrays so it only has to be done once", "# for linear interpola...
Advance particles using a gridded velocity field. Because of the Runga-Kutta timestepping, we need two velocity fields at different times. Parameters ---------- U0, V0 : array-like Gridded velocity fields at time t - dt. U1, V1 : array-like ...
[ "Advance", "particles", "using", "a", "gridded", "velocity", "field", ".", "Because", "of", "the", "Runga", "-", "Kutta", "timestepping", "we", "need", "two", "velocity", "fields", "at", "different", "times", ".", "Parameters", "----------", "U0", "V0", ":", ...
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L223-L258
pyqg/pyqg
pyqg/diagnostic_tools.py
spec_var
def spec_var(model, ph): """Compute variance of ``p`` from Fourier coefficients ``ph``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- var_dens :...
python
def spec_var(model, ph): """Compute variance of ``p`` from Fourier coefficients ``ph``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- var_dens :...
[ "def", "spec_var", "(", "model", ",", "ph", ")", ":", "var_dens", "=", "2.", "*", "np", ".", "abs", "(", "ph", ")", "**", "2", "/", "model", ".", "M", "**", "2", "# only half of coefs [0] and [nx/2+1] due to symmetry in real fft2", "var_dens", "[", "...", ...
Compute variance of ``p`` from Fourier coefficients ``ph``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- var_dens : float The variance of `...
[ "Compute", "variance", "of", "p", "from", "Fourier", "coefficients", "ph", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L7-L27
pyqg/pyqg
pyqg/diagnostic_tools.py
spec_sum
def spec_sum(ph2): """Compute total spectral sum of the real spectral quantity``ph^2``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph2 : real array The field on which to compute the sum Returns ------- var_dens : float ...
python
def spec_sum(ph2): """Compute total spectral sum of the real spectral quantity``ph^2``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph2 : real array The field on which to compute the sum Returns ------- var_dens : float ...
[ "def", "spec_sum", "(", "ph2", ")", ":", "ph2", "=", "2.", "*", "ph2", "ph2", "[", "...", ",", "0", "]", "=", "ph2", "[", "...", ",", "0", "]", "/", "2.", "ph2", "[", "...", ",", "-", "1", "]", "=", "ph2", "[", "...", ",", "-", "1", "]"...
Compute total spectral sum of the real spectral quantity``ph^2``. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph2 : real array The field on which to compute the sum Returns ------- var_dens : float The sum of `ph2`
[ "Compute", "total", "spectral", "sum", "of", "the", "real", "spectral", "quantity", "ph^2", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L30-L50
pyqg/pyqg
pyqg/diagnostic_tools.py
calc_ispec
def calc_ispec(model, ph): """Compute isotropic spectrum `phr` of `ph` from 2D spectrum. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- kr : arra...
python
def calc_ispec(model, ph): """Compute isotropic spectrum `phr` of `ph` from 2D spectrum. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- kr : arra...
[ "def", "calc_ispec", "(", "model", ",", "ph", ")", ":", "if", "model", ".", "kk", ".", "max", "(", ")", ">", "model", ".", "ll", ".", "max", "(", ")", ":", "kmax", "=", "model", ".", "ll", ".", "max", "(", ")", "else", ":", "kmax", "=", "mo...
Compute isotropic spectrum `phr` of `ph` from 2D spectrum. Parameters ---------- model : pyqg.Model instance The model object from which `ph` originates ph : complex array The field on which to compute the variance Returns ------- kr : array isotropic wavenumber ...
[ "Compute", "isotropic", "spectrum", "phr", "of", "ph", "from", "2D", "spectrum", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L53-L86
pyqg/pyqg
pyqg/layered_model.py
LayeredModel._initialize_stretching_matrix
def _initialize_stretching_matrix(self): """ Set up the stretching matrix """ self.S = np.zeros((self.nz, self.nz)) if (self.nz==2) and (self.rd) and (self.delta): self.del1 = self.delta/(self.delta+1.) self.del2 = (self.delta+1.)**-1 self.Us = self.Ubg[0]-...
python
def _initialize_stretching_matrix(self): """ Set up the stretching matrix """ self.S = np.zeros((self.nz, self.nz)) if (self.nz==2) and (self.rd) and (self.delta): self.del1 = self.delta/(self.delta+1.) self.del2 = (self.delta+1.)**-1 self.Us = self.Ubg[0]-...
[ "def", "_initialize_stretching_matrix", "(", "self", ")", ":", "self", ".", "S", "=", "np", ".", "zeros", "(", "(", "self", ".", "nz", ",", "self", ".", "nz", ")", ")", "if", "(", "self", ".", "nz", "==", "2", ")", "and", "(", "self", ".", "rd"...
Set up the stretching matrix
[ "Set", "up", "the", "stretching", "matrix" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L130-L162
pyqg/pyqg
pyqg/layered_model.py
LayeredModel._initialize_background
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" self.H = self.Hi.sum() if np.asarray(self.U).ndim == 2: self.Ubg = self.U * np.ones((self.ny)) else: self.Ubg = np.expand_dims(self.U,axis=1) * np.ones((self.ny)) ...
python
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" self.H = self.Hi.sum() if np.asarray(self.U).ndim == 2: self.Ubg = self.U * np.ones((self.ny)) else: self.Ubg = np.expand_dims(self.U,axis=1) * np.ones((self.ny)) ...
[ "def", "_initialize_background", "(", "self", ")", ":", "self", ".", "H", "=", "self", ".", "Hi", ".", "sum", "(", ")", "if", "np", ".", "asarray", "(", "self", ".", "U", ")", ".", "ndim", "==", "2", ":", "self", ".", "Ubg", "=", "self", ".", ...
Set up background state (zonal flow and PV gradients).
[ "Set", "up", "background", "state", "(", "zonal", "flow", "and", "PV", "gradients", ")", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L164-L208
pyqg/pyqg
pyqg/layered_model.py
LayeredModel._calc_eddy_time
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = 0. for j in range(self.nz): ens = .5*self.Hi[j] * self.spec_var(self.wv2*self.ph[j]) return 2.*pi*np.sqrt( self.H / ens.sum() ) / 86400
python
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = 0. for j in range(self.nz): ens = .5*self.Hi[j] * self.spec_var(self.wv2*self.ph[j]) return 2.*pi*np.sqrt( self.H / ens.sum() ) / 86400
[ "def", "_calc_eddy_time", "(", "self", ")", ":", "ens", "=", "0.", "for", "j", "in", "range", "(", "self", ".", "nz", ")", ":", "ens", "=", ".5", "*", "self", ".", "Hi", "[", "j", "]", "*", "self", ".", "spec_var", "(", "self", ".", "wv2", "*...
estimate the eddy turn-over time in days
[ "estimate", "the", "eddy", "turn", "-", "over", "time", "in", "days" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L255-L261
pyqg/pyqg
pyqg/layered_model.py
LayeredModel._initialize_model_diagnostics
def _initialize_model_diagnostics(self): """ Extra diagnostics for layered model """ self.add_diagnostic('entspec', description='barotropic enstrophy spectrum', function= (lambda self: np.abs((self.Hi[:,np.newaxis,np.newaxis]*self.qh).sum(axis=0))**2/...
python
def _initialize_model_diagnostics(self): """ Extra diagnostics for layered model """ self.add_diagnostic('entspec', description='barotropic enstrophy spectrum', function= (lambda self: np.abs((self.Hi[:,np.newaxis,np.newaxis]*self.qh).sum(axis=0))**2/...
[ "def", "_initialize_model_diagnostics", "(", "self", ")", ":", "self", ".", "add_diagnostic", "(", "'entspec'", ",", "description", "=", "'barotropic enstrophy spectrum'", ",", "function", "=", "(", "lambda", "self", ":", "np", ".", "abs", "(", "(", "self", "....
Extra diagnostics for layered model
[ "Extra", "diagnostics", "for", "layered", "model" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L277-L327
pyqg/pyqg
pyqg/qg_model.py
QGModel._initialize_background
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" # Background zonal flow (m/s): self.H = self.Hi.sum() self.set_U1U2(self.U1, self.U2) self.U = self.U1 - self.U2 # the F parameters self.F1 = self.rd**-2 / (1.+self.de...
python
def _initialize_background(self): """Set up background state (zonal flow and PV gradients).""" # Background zonal flow (m/s): self.H = self.Hi.sum() self.set_U1U2(self.U1, self.U2) self.U = self.U1 - self.U2 # the F parameters self.F1 = self.rd**-2 / (1.+self.de...
[ "def", "_initialize_background", "(", "self", ")", ":", "# Background zonal flow (m/s):", "self", ".", "H", "=", "self", ".", "Hi", ".", "sum", "(", ")", "self", ".", "set_U1U2", "(", "self", ".", "U1", ",", "self", ".", "U2", ")", "self", ".", "U", ...
Set up background state (zonal flow and PV gradients).
[ "Set", "up", "background", "state", "(", "zonal", "flow", "and", "PV", "gradients", ")", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L114-L142
pyqg/pyqg
pyqg/qg_model.py
QGModel.set_q1q2
def set_q1q2(self, q1, q2, check=False): """Set upper and lower layer PV anomalies. Parameters ---------- q1 : array-like Upper layer PV anomaly in spatial coordinates. q1 : array-like Lower layer PV anomaly in spatial coordinates. """ se...
python
def set_q1q2(self, q1, q2, check=False): """Set upper and lower layer PV anomalies. Parameters ---------- q1 : array-like Upper layer PV anomaly in spatial coordinates. q1 : array-like Lower layer PV anomaly in spatial coordinates. """ se...
[ "def", "set_q1q2", "(", "self", ",", "q1", ",", "q2", ",", "check", "=", "False", ")", ":", "self", ".", "set_q", "(", "np", ".", "vstack", "(", "[", "q1", "[", "np", ".", "newaxis", ",", ":", ",", ":", "]", ",", "q2", "[", "np", ".", "newa...
Set upper and lower layer PV anomalies. Parameters ---------- q1 : array-like Upper layer PV anomaly in spatial coordinates. q1 : array-like Lower layer PV anomaly in spatial coordinates.
[ "Set", "upper", "and", "lower", "layer", "PV", "anomalies", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L170-L191
pyqg/pyqg
pyqg/qg_model.py
QGModel.set_U1U2
def set_U1U2(self, U1, U2): """Set background zonal flow. Parameters ---------- U1 : number Upper layer flow. Units: m/s U2 : number Lower layer flow. Units: m/s """ if len(np.shape(U1)) == 0: U1 = U1 * np.ones((self.ny)) ...
python
def set_U1U2(self, U1, U2): """Set background zonal flow. Parameters ---------- U1 : number Upper layer flow. Units: m/s U2 : number Lower layer flow. Units: m/s """ if len(np.shape(U1)) == 0: U1 = U1 * np.ones((self.ny)) ...
[ "def", "set_U1U2", "(", "self", ",", "U1", ",", "U2", ")", ":", "if", "len", "(", "np", ".", "shape", "(", "U1", ")", ")", "==", "0", ":", "U1", "=", "U1", "*", "np", ".", "ones", "(", "(", "self", ".", "ny", ")", ")", "if", "len", "(", ...
Set background zonal flow. Parameters ---------- U1 : number Upper layer flow. Units: m/s U2 : number Lower layer flow. Units: m/s
[ "Set", "background", "zonal", "flow", "." ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L193-L211
pyqg/pyqg
pyqg/qg_model.py
QGModel._calc_eddy_time
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = .5*self.Hi[0] * self.spec_var(self.wv2*self.ph1) + \ .5*self.Hi[1] * self.spec_var(self.wv2*self.ph2) return 2.*pi*np.sqrt( self.H / ens ) / 86400
python
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = .5*self.Hi[0] * self.spec_var(self.wv2*self.ph1) + \ .5*self.Hi[1] * self.spec_var(self.wv2*self.ph2) return 2.*pi*np.sqrt( self.H / ens ) / 86400
[ "def", "_calc_eddy_time", "(", "self", ")", ":", "ens", "=", ".5", "*", "self", ".", "Hi", "[", "0", "]", "*", "self", ".", "spec_var", "(", "self", ".", "wv2", "*", "self", ".", "ph1", ")", "+", ".5", "*", "self", ".", "Hi", "[", "1", "]", ...
estimate the eddy turn-over time in days
[ "estimate", "the", "eddy", "turn", "-", "over", "time", "in", "days" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L228-L234
pyqg/pyqg
pyqg/qg_model.py
QGModel._initialize_model_diagnostics
def _initialize_model_diagnostics(self): """Extra diagnostics for two-layer model""" self.add_diagnostic('entspec', description='barotropic enstrophy spectrum', function= (lambda self: np.abs(self.del1*self.qh[0] + self.del2*self.qh[1])**2.) ) ...
python
def _initialize_model_diagnostics(self): """Extra diagnostics for two-layer model""" self.add_diagnostic('entspec', description='barotropic enstrophy spectrum', function= (lambda self: np.abs(self.del1*self.qh[0] + self.del2*self.qh[1])**2.) ) ...
[ "def", "_initialize_model_diagnostics", "(", "self", ")", ":", "self", ".", "add_diagnostic", "(", "'entspec'", ",", "description", "=", "'barotropic enstrophy spectrum'", ",", "function", "=", "(", "lambda", "self", ":", "np", ".", "abs", "(", "self", ".", "d...
Extra diagnostics for two-layer model
[ "Extra", "diagnostics", "for", "two", "-", "layer", "model" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L246-L286
pyqg/pyqg
pyqg/bt_model.py
BTModel._initialize_inversion_matrix
def _initialize_inversion_matrix(self): """ the inversion """ # The bt model is diagonal. The inversion is simply qh = -kappa**2 ph self.a = -(self.wv2i+self.kd2)[np.newaxis, np.newaxis, :, :]
python
def _initialize_inversion_matrix(self): """ the inversion """ # The bt model is diagonal. The inversion is simply qh = -kappa**2 ph self.a = -(self.wv2i+self.kd2)[np.newaxis, np.newaxis, :, :]
[ "def", "_initialize_inversion_matrix", "(", "self", ")", ":", "# The bt model is diagonal. The inversion is simply qh = -kappa**2 ph", "self", ".", "a", "=", "-", "(", "self", ".", "wv2i", "+", "self", ".", "kd2", ")", "[", "np", ".", "newaxis", ",", "np", ".", ...
the inversion
[ "the", "inversion" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/bt_model.py#L75-L78
pyqg/pyqg
pyqg/bt_model.py
BTModel._calc_eddy_time
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = .5*self.H * self.spec_var(self.wv2*self.ph) return 2.*pi*np.sqrt( self.H / ens ) / year
python
def _calc_eddy_time(self): """ estimate the eddy turn-over time in days """ ens = .5*self.H * self.spec_var(self.wv2*self.ph) return 2.*pi*np.sqrt( self.H / ens ) / year
[ "def", "_calc_eddy_time", "(", "self", ")", ":", "ens", "=", ".5", "*", "self", ".", "H", "*", "self", ".", "spec_var", "(", "self", ".", "wv2", "*", "self", ".", "ph", ")", "return", "2.", "*", "pi", "*", "np", ".", "sqrt", "(", "self", ".", ...
estimate the eddy turn-over time in days
[ "estimate", "the", "eddy", "turn", "-", "over", "time", "in", "days" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/bt_model.py#L123-L126
pyqg/pyqg
pyqg/point_vortex.py
PointVortexArray2D.calc_uv
def calc_uv(self, x, y, prev=False): """Calculate velocity at x and y points due to vortex velocity field. Assumes x and y are vortex positions and are ordered the same as x0 and y0. The ordering is used to neglect to vortex self interaction.""" assert len(x) == self.N assert len...
python
def calc_uv(self, x, y, prev=False): """Calculate velocity at x and y points due to vortex velocity field. Assumes x and y are vortex positions and are ordered the same as x0 and y0. The ordering is used to neglect to vortex self interaction.""" assert len(x) == self.N assert len...
[ "def", "calc_uv", "(", "self", ",", "x", ",", "y", ",", "prev", "=", "False", ")", ":", "assert", "len", "(", "x", ")", "==", "self", ".", "N", "assert", "len", "(", "y", ")", "==", "self", ".", "N", "u", "=", "np", ".", "zeros", "(", "self...
Calculate velocity at x and y points due to vortex velocity field. Assumes x and y are vortex positions and are ordered the same as x0 and y0. The ordering is used to neglect to vortex self interaction.
[ "Calculate", "velocity", "at", "x", "and", "y", "points", "due", "to", "vortex", "velocity", "field", ".", "Assumes", "x", "and", "y", "are", "vortex", "positions", "and", "are", "ordered", "the", "same", "as", "x0", "and", "y0", ".", "The", "ordering", ...
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/point_vortex.py#L32-L52
pyqg/pyqg
pyqg/point_vortex.py
PointVortexArray2D.uv_at_xy
def uv_at_xy(self, x, y, x0, y0, s0): """Returns two arrays of u, v""" dx, dy = self.distance(x0, y0, x, y) #print 'dx, dy:', dx, dy rr2 = (dx**2 + dy**2)**-1 u = - s0 * dy * r_twopi * rr2 v = s0 * dx * r_twopi * rr2 #print 'u, v', u, v return u, v
python
def uv_at_xy(self, x, y, x0, y0, s0): """Returns two arrays of u, v""" dx, dy = self.distance(x0, y0, x, y) #print 'dx, dy:', dx, dy rr2 = (dx**2 + dy**2)**-1 u = - s0 * dy * r_twopi * rr2 v = s0 * dx * r_twopi * rr2 #print 'u, v', u, v return u, v
[ "def", "uv_at_xy", "(", "self", ",", "x", ",", "y", ",", "x0", ",", "y0", ",", "s0", ")", ":", "dx", ",", "dy", "=", "self", ".", "distance", "(", "x0", ",", "y0", ",", "x", ",", "y", ")", "#print 'dx, dy:', dx, dy", "rr2", "=", "(", "dx", "*...
Returns two arrays of u, v
[ "Returns", "two", "arrays", "of", "u", "v" ]
train
https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/point_vortex.py#L54-L62
brentp/interlap
interlap.py
reduce
def reduce(args): """ >>> reduce([(2, 4), (4, 9)]) [(2, 4), (4, 9)] >>> reduce([(2, 6), (4, 10)]) [(2, 10)] """ if len(args) < 2: return args args.sort() ret = [args[0]] for next_i, (s, e) in enumerate(args, start=1): if next_i == len(args): ret[-1] = ret[-1]...
python
def reduce(args): """ >>> reduce([(2, 4), (4, 9)]) [(2, 4), (4, 9)] >>> reduce([(2, 6), (4, 10)]) [(2, 10)] """ if len(args) < 2: return args args.sort() ret = [args[0]] for next_i, (s, e) in enumerate(args, start=1): if next_i == len(args): ret[-1] = ret[-1]...
[ "def", "reduce", "(", "args", ")", ":", "if", "len", "(", "args", ")", "<", "2", ":", "return", "args", "args", ".", "sort", "(", ")", "ret", "=", "[", "args", "[", "0", "]", "]", "for", "next_i", ",", "(", "s", ",", "e", ")", "in", "enumer...
>>> reduce([(2, 4), (4, 9)]) [(2, 4), (4, 9)] >>> reduce([(2, 6), (4, 10)]) [(2, 10)]
[ ">>>", "reduce", "(", "[", "(", "2", "4", ")", "(", "4", "9", ")", "]", ")", "[", "(", "2", "4", ")", "(", "4", "9", ")", "]" ]
train
https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L224-L245
brentp/interlap
interlap.py
InterLap.add
def add(self, ranges): r"""Add a single (or many) [start, end, \*] item to the tree.""" if len(ranges) and isinstance(ranges[0], int_types): ranges = [ranges] iset = self._iset self._maxlen = max(self._maxlen, max(r[1] - r[0] + 1 for r in ranges)) if len(ranges) > 30...
python
def add(self, ranges): r"""Add a single (or many) [start, end, \*] item to the tree.""" if len(ranges) and isinstance(ranges[0], int_types): ranges = [ranges] iset = self._iset self._maxlen = max(self._maxlen, max(r[1] - r[0] + 1 for r in ranges)) if len(ranges) > 30...
[ "def", "add", "(", "self", ",", "ranges", ")", ":", "if", "len", "(", "ranges", ")", "and", "isinstance", "(", "ranges", "[", "0", "]", ",", "int_types", ")", ":", "ranges", "=", "[", "ranges", "]", "iset", "=", "self", ".", "_iset", "self", ".",...
r"""Add a single (or many) [start, end, \*] item to the tree.
[ "r", "Add", "a", "single", "(", "or", "many", ")", "[", "start", "end", "\\", "*", "]", "item", "to", "the", "tree", "." ]
train
https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L133-L145
brentp/interlap
interlap.py
InterLap.find
def find(self, other): """Return an interable of elements that overlap other in the tree.""" iset = self._iset l = binsearch_left_start(iset, other[0] - self._maxlen, 0, len(iset)) r = binsearch_right_end(iset, other[1], 0, len(iset)) iopts = iset[l:r] iiter = (s for s in...
python
def find(self, other): """Return an interable of elements that overlap other in the tree.""" iset = self._iset l = binsearch_left_start(iset, other[0] - self._maxlen, 0, len(iset)) r = binsearch_right_end(iset, other[1], 0, len(iset)) iopts = iset[l:r] iiter = (s for s in...
[ "def", "find", "(", "self", ",", "other", ")", ":", "iset", "=", "self", ".", "_iset", "l", "=", "binsearch_left_start", "(", "iset", ",", "other", "[", "0", "]", "-", "self", ".", "_maxlen", ",", "0", ",", "len", "(", "iset", ")", ")", "r", "=...
Return an interable of elements that overlap other in the tree.
[ "Return", "an", "interable", "of", "elements", "that", "overlap", "other", "in", "the", "tree", "." ]
train
https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L153-L160
gumblex/zhconv
zhconv/zhconv.py
loaddict
def loaddict(filename=DICTIONARY): """ Load the dictionary from a specific JSON file. """ global zhcdicts if zhcdicts: return if filename == _DEFAULT_DICT: zhcdicts = json.loads(get_module_res(filename).read().decode('utf-8')) else: with open(filename, 'rb') as f: ...
python
def loaddict(filename=DICTIONARY): """ Load the dictionary from a specific JSON file. """ global zhcdicts if zhcdicts: return if filename == _DEFAULT_DICT: zhcdicts = json.loads(get_module_res(filename).read().decode('utf-8')) else: with open(filename, 'rb') as f: ...
[ "def", "loaddict", "(", "filename", "=", "DICTIONARY", ")", ":", "global", "zhcdicts", "if", "zhcdicts", ":", "return", "if", "filename", "==", "_DEFAULT_DICT", ":", "zhcdicts", "=", "json", ".", "loads", "(", "get_module_res", "(", "filename", ")", ".", "...
Load the dictionary from a specific JSON file.
[ "Load", "the", "dictionary", "from", "a", "specific", "JSON", "file", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L68-L81
gumblex/zhconv
zhconv/zhconv.py
getdict
def getdict(locale): """ Generate or get convertion dict cache for certain locale. Dictionaries are loaded on demand. """ global zhcdicts, dict_zhcn, dict_zhsg, dict_zhtw, dict_zhhk, pfsdict if zhcdicts is None: loaddict(DICTIONARY) if locale == 'zh-cn': if dict_zhcn: ...
python
def getdict(locale): """ Generate or get convertion dict cache for certain locale. Dictionaries are loaded on demand. """ global zhcdicts, dict_zhcn, dict_zhsg, dict_zhtw, dict_zhhk, pfsdict if zhcdicts is None: loaddict(DICTIONARY) if locale == 'zh-cn': if dict_zhcn: ...
[ "def", "getdict", "(", "locale", ")", ":", "global", "zhcdicts", ",", "dict_zhcn", ",", "dict_zhsg", ",", "dict_zhtw", ",", "dict_zhhk", ",", "pfsdict", "if", "zhcdicts", "is", "None", ":", "loaddict", "(", "DICTIONARY", ")", "if", "locale", "==", "'zh-cn'...
Generate or get convertion dict cache for certain locale. Dictionaries are loaded on demand.
[ "Generate", "or", "get", "convertion", "dict", "cache", "for", "certain", "locale", ".", "Dictionaries", "are", "loaded", "on", "demand", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L83-L127
gumblex/zhconv
zhconv/zhconv.py
issimp
def issimp(s, full=False): """ Detect text is whether Simplified Chinese or Traditional Chinese. Returns True for Simplified; False for Traditional; None for unknown. If full=False, it returns once first simplified- or traditional-only character is encountered, so it's for quick and rough identifica...
python
def issimp(s, full=False): """ Detect text is whether Simplified Chinese or Traditional Chinese. Returns True for Simplified; False for Traditional; None for unknown. If full=False, it returns once first simplified- or traditional-only character is encountered, so it's for quick and rough identifica...
[ "def", "issimp", "(", "s", ",", "full", "=", "False", ")", ":", "if", "zhcdicts", "is", "None", ":", "loaddict", "(", "DICTIONARY", ")", "simp", ",", "trad", "=", "0", ",", "0", "if", "full", ":", "for", "ch", "in", "s", ":", "if", "ch", "in", ...
Detect text is whether Simplified Chinese or Traditional Chinese. Returns True for Simplified; False for Traditional; None for unknown. If full=False, it returns once first simplified- or traditional-only character is encountered, so it's for quick and rough identification; else, it compares the count a...
[ "Detect", "text", "is", "whether", "Simplified", "Chinese", "or", "Traditional", "Chinese", ".", "Returns", "True", "for", "Simplified", ";", "False", "for", "Traditional", ";", "None", "for", "unknown", ".", "If", "full", "=", "False", "it", "returns", "onc...
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L136-L168
gumblex/zhconv
zhconv/zhconv.py
convtable2dict
def convtable2dict(convtable, locale, update=None): """ Convert a list of conversion dict to a dict for a certain locale. >>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items()) [('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('...
python
def convtable2dict(convtable, locale, update=None): """ Convert a list of conversion dict to a dict for a certain locale. >>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items()) [('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('...
[ "def", "convtable2dict", "(", "convtable", ",", "locale", ",", "update", "=", "None", ")", ":", "rdict", "=", "update", ".", "copy", "(", ")", "if", "update", "else", "{", "}", "for", "r", "in", "convtable", ":", "if", "':uni'", "in", "r", ":", "if...
Convert a list of conversion dict to a dict for a certain locale. >>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items()) [('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('里茲', '利兹')]
[ "Convert", "a", "list", "of", "conversion", "dict", "to", "a", "dict", "for", "a", "certain", "locale", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L176-L196
gumblex/zhconv
zhconv/zhconv.py
tokenize
def tokenize(s, locale, update=None): """ Tokenize `s` according to corresponding locale dictionary. Don't use this for serious text processing. """ zhdict = getdict(locale) pfset = pfsdict[locale] if update: zhdict = zhdict.copy() zhdict.update(update) newset = set()...
python
def tokenize(s, locale, update=None): """ Tokenize `s` according to corresponding locale dictionary. Don't use this for serious text processing. """ zhdict = getdict(locale) pfset = pfsdict[locale] if update: zhdict = zhdict.copy() zhdict.update(update) newset = set()...
[ "def", "tokenize", "(", "s", ",", "locale", ",", "update", "=", "None", ")", ":", "zhdict", "=", "getdict", "(", "locale", ")", "pfset", "=", "pfsdict", "[", "locale", "]", "if", "update", ":", "zhdict", "=", "zhdict", ".", "copy", "(", ")", "zhdic...
Tokenize `s` according to corresponding locale dictionary. Don't use this for serious text processing.
[ "Tokenize", "s", "according", "to", "corresponding", "locale", "dictionary", ".", "Don", "t", "use", "this", "for", "serious", "text", "processing", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L198-L233
gumblex/zhconv
zhconv/zhconv.py
convert_for_mw
def convert_for_mw(s, locale, update=None): """ Recognizes MediaWiki's human conversion format. Use locale='zh' for no conversion. Reference: (all tests passed) https://zh.wikipedia.org/wiki/Help:高级字词转换语法 https://www.mediawiki.org/wiki/Writing_systems/Syntax >>> print(convert_for_mw('在现代,机...
python
def convert_for_mw(s, locale, update=None): """ Recognizes MediaWiki's human conversion format. Use locale='zh' for no conversion. Reference: (all tests passed) https://zh.wikipedia.org/wiki/Help:高级字词转换语法 https://www.mediawiki.org/wiki/Writing_systems/Syntax >>> print(convert_for_mw('在现代,机...
[ "def", "convert_for_mw", "(", "s", ",", "locale", ",", "update", "=", "None", ")", ":", "ch", "=", "[", "]", "rules", "=", "[", "]", "ruledict", "=", "update", ".", "copy", "(", ")", "if", "update", "else", "{", "}", "nested", "=", "0", "block", ...
Recognizes MediaWiki's human conversion format. Use locale='zh' for no conversion. Reference: (all tests passed) https://zh.wikipedia.org/wiki/Help:高级字词转换语法 https://www.mediawiki.org/wiki/Writing_systems/Syntax >>> print(convert_for_mw('在现代,机械计算-{}-机的应用已经完全被电子计算-{}-机所取代', 'zh-hk')) 在現代,機械計算機的應...
[ "Recognizes", "MediaWiki", "s", "human", "conversion", "format", ".", "Use", "locale", "=", "zh", "for", "no", "conversion", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L292-L425
gumblex/zhconv
zhconv/zhconv.py
main
def main(): """ Simple stdin/stdout interface. """ if len(sys.argv) == 2 and sys.argv[1] in Locales: locale = sys.argv[1] convertfunc = convert elif len(sys.argv) == 3 and sys.argv[1] == '-w' and sys.argv[2] in Locales: locale = sys.argv[2] convertfunc = convert_for_m...
python
def main(): """ Simple stdin/stdout interface. """ if len(sys.argv) == 2 and sys.argv[1] in Locales: locale = sys.argv[1] convertfunc = convert elif len(sys.argv) == 3 and sys.argv[1] == '-w' and sys.argv[2] in Locales: locale = sys.argv[2] convertfunc = convert_for_m...
[ "def", "main", "(", ")", ":", "if", "len", "(", "sys", ".", "argv", ")", "==", "2", "and", "sys", ".", "argv", "[", "1", "]", "in", "Locales", ":", "locale", "=", "sys", ".", "argv", "[", "1", "]", "convertfunc", "=", "convert", "elif", "len", ...
Simple stdin/stdout interface.
[ "Simple", "stdin", "/", "stdout", "interface", "." ]
train
https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L449-L475
glasslion/django-qiniu-storage
qiniustorage/utils.py
bucket_lister
def bucket_lister(manager, bucket_name, prefix=None, marker=None, limit=None): """ A generator function for listing keys in a bucket. """ eof = False while not eof: ret, eof, info = manager.list(bucket_name, prefix=prefix, limit=limit, marker=marker) ...
python
def bucket_lister(manager, bucket_name, prefix=None, marker=None, limit=None): """ A generator function for listing keys in a bucket. """ eof = False while not eof: ret, eof, info = manager.list(bucket_name, prefix=prefix, limit=limit, marker=marker) ...
[ "def", "bucket_lister", "(", "manager", ",", "bucket_name", ",", "prefix", "=", "None", ",", "marker", "=", "None", ",", "limit", "=", "None", ")", ":", "eof", "=", "False", "while", "not", "eof", ":", "ret", ",", "eof", ",", "info", "=", "manager", ...
A generator function for listing keys in a bucket.
[ "A", "generator", "function", "for", "listing", "keys", "in", "a", "bucket", "." ]
train
https://github.com/glasslion/django-qiniu-storage/blob/b046ec0b67ebcf8cd9eb09c60f7db4a7e4fab7ad/qiniustorage/utils.py#L17-L31
glasslion/django-qiniu-storage
qiniustorage/backends.py
get_qiniu_config
def get_qiniu_config(name, default=None): """ Get configuration variable from environment variable or django setting.py """ config = os.environ.get(name, getattr(settings, name, default)) if config is not None: if isinstance(config, six.string_types): return config.strip() ...
python
def get_qiniu_config(name, default=None): """ Get configuration variable from environment variable or django setting.py """ config = os.environ.get(name, getattr(settings, name, default)) if config is not None: if isinstance(config, six.string_types): return config.strip() ...
[ "def", "get_qiniu_config", "(", "name", ",", "default", "=", "None", ")", ":", "config", "=", "os", ".", "environ", ".", "get", "(", "name", ",", "getattr", "(", "settings", ",", "name", ",", "default", ")", ")", "if", "config", "is", "not", "None", ...
Get configuration variable from environment variable or django setting.py
[ "Get", "configuration", "variable", "from", "environment", "variable", "or", "django", "setting", ".", "py" ]
train
https://github.com/glasslion/django-qiniu-storage/blob/b046ec0b67ebcf8cd9eb09c60f7db4a7e4fab7ad/qiniustorage/backends.py#L27-L41
non-Jedi/gyr
gyr/api.py
MatrixASHttpAPI.register
def register(self, username=""): """Performs /register with type: m.login.application_service Args: username(str): Username to register. """ if not username: username = utils.mxid2localpart(self.identity) content = { "type": "m.login.applicati...
python
def register(self, username=""): """Performs /register with type: m.login.application_service Args: username(str): Username to register. """ if not username: username = utils.mxid2localpart(self.identity) content = { "type": "m.login.applicati...
[ "def", "register", "(", "self", ",", "username", "=", "\"\"", ")", ":", "if", "not", "username", ":", "username", "=", "utils", ".", "mxid2localpart", "(", "self", ".", "identity", ")", "content", "=", "{", "\"type\"", ":", "\"m.login.application_service\"",...
Performs /register with type: m.login.application_service Args: username(str): Username to register.
[ "Performs", "/", "register", "with", "type", ":", "m", ".", "login", ".", "application_service" ]
train
https://github.com/non-Jedi/gyr/blob/9f7bfe033b9d3bbfd3a9e8aea02e35526b53125e/gyr/api.py#L58-L71
ocaballeror/LyricFetch
lyricfetch/cli.py
load_from_file
def load_from_file(filename): """ Load a list of filenames from an external text file. """ if os.path.isdir(filename): logger.error("Err: File '%s' is a directory", filename) return None if not os.path.isfile(filename): logger.error("Err: File '%s' does not exist", filename) ...
python
def load_from_file(filename): """ Load a list of filenames from an external text file. """ if os.path.isdir(filename): logger.error("Err: File '%s' is a directory", filename) return None if not os.path.isfile(filename): logger.error("Err: File '%s' does not exist", filename) ...
[ "def", "load_from_file", "(", "filename", ")", ":", "if", "os", ".", "path", ".", "isdir", "(", "filename", ")", ":", "logger", ".", "error", "(", "\"Err: File '%s' is a directory\"", ",", "filename", ")", "return", "None", "if", "not", "os", ".", "path", ...
Load a list of filenames from an external text file.
[ "Load", "a", "list", "of", "filenames", "from", "an", "external", "text", "file", "." ]
train
https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L17-L35
ocaballeror/LyricFetch
lyricfetch/cli.py
parse_argv
def parse_argv(): """ Parse command line arguments. Settings will be stored in the global variables declared above. """ parser = argparse.ArgumentParser(description='Find lyrics for a set of mp3' ' files and embed them as metadata') parser.add_argument('-j', ...
python
def parse_argv(): """ Parse command line arguments. Settings will be stored in the global variables declared above. """ parser = argparse.ArgumentParser(description='Find lyrics for a set of mp3' ' files and embed them as metadata') parser.add_argument('-j', ...
[ "def", "parse_argv", "(", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", "description", "=", "'Find lyrics for a set of mp3'", "' files and embed them as metadata'", ")", "parser", ".", "add_argument", "(", "'-j'", ",", "'--jobs'", ",", "help", "="...
Parse command line arguments. Settings will be stored in the global variables declared above.
[ "Parse", "command", "line", "arguments", ".", "Settings", "will", "be", "stored", "in", "the", "global", "variables", "declared", "above", "." ]
train
https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L38-L99
ocaballeror/LyricFetch
lyricfetch/cli.py
main
def main(): """ Main function. """ msg = '' try: songs = parse_argv() if not songs: msg = 'No songs specified' except ValueError as error: msg = str(error) if msg: logger.error('%s: Error: %s', sys.argv[0], msg) return 1 logger.debug('...
python
def main(): """ Main function. """ msg = '' try: songs = parse_argv() if not songs: msg = 'No songs specified' except ValueError as error: msg = str(error) if msg: logger.error('%s: Error: %s', sys.argv[0], msg) return 1 logger.debug('...
[ "def", "main", "(", ")", ":", "msg", "=", "''", "try", ":", "songs", "=", "parse_argv", "(", ")", "if", "not", "songs", ":", "msg", "=", "'No songs specified'", "except", "ValueError", "as", "error", ":", "msg", "=", "str", "(", "error", ")", "if", ...
Main function.
[ "Main", "function", "." ]
train
https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L102-L124
taskcluster/slugid.py
slugid/slugid.py
decode
def decode(slug): """ Returns the uuid.UUID object represented by the given v4 or "nice" slug """ if sys.version_info.major != 2 and isinstance(slug, bytes): slug = slug.decode('ascii') slug = slug + '==' # base64 padding return uuid.UUID(bytes=base64.urlsafe_b64decode(slug))
python
def decode(slug): """ Returns the uuid.UUID object represented by the given v4 or "nice" slug """ if sys.version_info.major != 2 and isinstance(slug, bytes): slug = slug.decode('ascii') slug = slug + '==' # base64 padding return uuid.UUID(bytes=base64.urlsafe_b64decode(slug))
[ "def", "decode", "(", "slug", ")", ":", "if", "sys", ".", "version_info", ".", "major", "!=", "2", "and", "isinstance", "(", "slug", ",", "bytes", ")", ":", "slug", "=", "slug", ".", "decode", "(", "'ascii'", ")", "slug", "=", "slug", "+", "'=='", ...
Returns the uuid.UUID object represented by the given v4 or "nice" slug
[ "Returns", "the", "uuid", ".", "UUID", "object", "represented", "by", "the", "given", "v4", "or", "nice", "slug" ]
train
https://github.com/taskcluster/slugid.py/blob/7c2c58e79d8684a54c578302ad60b384e52bb09b/slugid/slugid.py#L24-L31
taskcluster/slugid.py
slugid/slugid.py
nice
def nice(): """ Returns a randomly generated uuid v4 compliant slug which conforms to a set of "nice" properties, at the cost of some entropy. Currently this means one extra fixed bit (the first bit of the uuid is set to 0) which guarantees the slug will begin with [A-Za-f]. For example such slugs d...
python
def nice(): """ Returns a randomly generated uuid v4 compliant slug which conforms to a set of "nice" properties, at the cost of some entropy. Currently this means one extra fixed bit (the first bit of the uuid is set to 0) which guarantees the slug will begin with [A-Za-f]. For example such slugs d...
[ "def", "nice", "(", ")", ":", "rawBytes", "=", "bytearray", "(", "uuid", ".", "uuid4", "(", ")", ".", "bytes", ")", "rawBytes", "[", "0", "]", "=", "rawBytes", "[", "0", "]", "&", "0x7f", "# Ensure slug starts with [A-Za-f]", "return", "_convert_bytes_to_s...
Returns a randomly generated uuid v4 compliant slug which conforms to a set of "nice" properties, at the cost of some entropy. Currently this means one extra fixed bit (the first bit of the uuid is set to 0) which guarantees the slug will begin with [A-Za-f]. For example such slugs don't require special ...
[ "Returns", "a", "randomly", "generated", "uuid", "v4", "compliant", "slug", "which", "conforms", "to", "a", "set", "of", "nice", "properties", "at", "the", "cost", "of", "some", "entropy", ".", "Currently", "this", "means", "one", "extra", "fixed", "bit", ...
train
https://github.com/taskcluster/slugid.py/blob/7c2c58e79d8684a54c578302ad60b384e52bb09b/slugid/slugid.py#L41-L55
inodb/sufam
sufam/mutation.py
MutationsAtSinglePosition.filter_against_normal
def filter_against_normal(self, normal_mutations, maf_min=0.2, maf_count_threshold=20, count_min=1): """Filters mutations that are in the given normal""" assert(normal_mutations.chrom == self.chrom) assert(normal_mutations.pos == self.pos) assert(normal_muta...
python
def filter_against_normal(self, normal_mutations, maf_min=0.2, maf_count_threshold=20, count_min=1): """Filters mutations that are in the given normal""" assert(normal_mutations.chrom == self.chrom) assert(normal_mutations.pos == self.pos) assert(normal_muta...
[ "def", "filter_against_normal", "(", "self", ",", "normal_mutations", ",", "maf_min", "=", "0.2", ",", "maf_count_threshold", "=", "20", ",", "count_min", "=", "1", ")", ":", "assert", "(", "normal_mutations", ".", "chrom", "==", "self", ".", "chrom", ")", ...
Filters mutations that are in the given normal
[ "Filters", "mutations", "that", "are", "in", "the", "given", "normal" ]
train
https://github.com/inodb/sufam/blob/d4e41c5478ca9ba58be44d95106885c096c90a74/sufam/mutation.py#L55-L81
inodb/sufam
sufam/mutation.py
Mutation.to_oncotator
def to_oncotator(self): """Returns mutation in oncotator input format. Assumes mutations have vcf/mpileup style positions.""" if self.type == ".": ref = self.ref alt = self.change start = self.pos end = self.pos elif self.type == "-": ...
python
def to_oncotator(self): """Returns mutation in oncotator input format. Assumes mutations have vcf/mpileup style positions.""" if self.type == ".": ref = self.ref alt = self.change start = self.pos end = self.pos elif self.type == "-": ...
[ "def", "to_oncotator", "(", "self", ")", ":", "if", "self", ".", "type", "==", "\".\"", ":", "ref", "=", "self", ".", "ref", "alt", "=", "self", ".", "change", "start", "=", "self", ".", "pos", "end", "=", "self", ".", "pos", "elif", "self", ".",...
Returns mutation in oncotator input format. Assumes mutations have vcf/mpileup style positions.
[ "Returns", "mutation", "in", "oncotator", "input", "format", ".", "Assumes", "mutations", "have", "vcf", "/", "mpileup", "style", "positions", "." ]
train
https://github.com/inodb/sufam/blob/d4e41c5478ca9ba58be44d95106885c096c90a74/sufam/mutation.py#L116-L137
non-Jedi/gyr
gyr/server.py
Application.add_handlers
def add_handlers(self, room_handler=None, transaction_handler=None, user_handler=None): """Adds routes to Application that use specified handlers.""" # Add all the normal matrix API routes if room_handler: room = resources.Room(room_handler, ...
python
def add_handlers(self, room_handler=None, transaction_handler=None, user_handler=None): """Adds routes to Application that use specified handlers.""" # Add all the normal matrix API routes if room_handler: room = resources.Room(room_handler, ...
[ "def", "add_handlers", "(", "self", ",", "room_handler", "=", "None", ",", "transaction_handler", "=", "None", ",", "user_handler", "=", "None", ")", ":", "# Add all the normal matrix API routes", "if", "room_handler", ":", "room", "=", "resources", ".", "Room", ...
Adds routes to Application that use specified handlers.
[ "Adds", "routes", "to", "Application", "that", "use", "specified", "handlers", "." ]
train
https://github.com/non-Jedi/gyr/blob/9f7bfe033b9d3bbfd3a9e8aea02e35526b53125e/gyr/server.py#L34-L49
HolmesNL/confidence
confidence/utils.py
_merge
def _merge(left, right, path=None, conflict=_Conflict.error): """ Merges values in place from *right* into *left*. :param left: mapping to merge into :param right: mapping to merge from :param path: `list` of keys processed before (used for error reporting only, should only need to be provi...
python
def _merge(left, right, path=None, conflict=_Conflict.error): """ Merges values in place from *right* into *left*. :param left: mapping to merge into :param right: mapping to merge from :param path: `list` of keys processed before (used for error reporting only, should only need to be provi...
[ "def", "_merge", "(", "left", ",", "right", ",", "path", "=", "None", ",", "conflict", "=", "_Conflict", ".", "error", ")", ":", "path", "=", "path", "or", "[", "]", "conflict", "=", "_Conflict", "(", "conflict", ")", "for", "key", "in", "right", "...
Merges values in place from *right* into *left*. :param left: mapping to merge into :param right: mapping to merge from :param path: `list` of keys processed before (used for error reporting only, should only need to be provided by recursive calls) :param conflict: action to be taken on merge c...
[ "Merges", "values", "in", "place", "from", "*", "right", "*", "into", "*", "left", "*", "." ]
train
https://github.com/HolmesNL/confidence/blob/e14d2d8769a01fa55676716f7a2f22714c2616d3/confidence/utils.py#L13-L46
HolmesNL/confidence
confidence/utils.py
_split_keys
def _split_keys(mapping, separator='.', colliding=None): """ Recursively walks *mapping* to split keys that contain the separator into nested mappings. .. note:: Keys not of type `str` are not supported and will raise errors. :param mapping: the mapping to process :param separator: th...
python
def _split_keys(mapping, separator='.', colliding=None): """ Recursively walks *mapping* to split keys that contain the separator into nested mappings. .. note:: Keys not of type `str` are not supported and will raise errors. :param mapping: the mapping to process :param separator: th...
[ "def", "_split_keys", "(", "mapping", ",", "separator", "=", "'.'", ",", "colliding", "=", "None", ")", ":", "result", "=", "{", "}", "for", "key", ",", "value", "in", "mapping", ".", "items", "(", ")", ":", "if", "isinstance", "(", "value", ",", "...
Recursively walks *mapping* to split keys that contain the separator into nested mappings. .. note:: Keys not of type `str` are not supported and will raise errors. :param mapping: the mapping to process :param separator: the character (sequence) to use as the separator between keys ...
[ "Recursively", "walks", "*", "mapping", "*", "to", "split", "keys", "that", "contain", "the", "separator", "into", "nested", "mappings", "." ]
train
https://github.com/HolmesNL/confidence/blob/e14d2d8769a01fa55676716f7a2f22714c2616d3/confidence/utils.py#L49-L91
tipsi/tipsi_tools
tipsi_tools/monitoring.py
log_mon_value
def log_mon_value(name, value=1, **kwargs): """ simplest monitoring function to be aggregated with sum """ message = '{} => {}'.format(name, value) log_mon.info({'metric_name': name, 'value': value, 'message': message, **kwargs})
python
def log_mon_value(name, value=1, **kwargs): """ simplest monitoring function to be aggregated with sum """ message = '{} => {}'.format(name, value) log_mon.info({'metric_name': name, 'value': value, 'message': message, **kwargs})
[ "def", "log_mon_value", "(", "name", ",", "value", "=", "1", ",", "*", "*", "kwargs", ")", ":", "message", "=", "'{} => {}'", ".", "format", "(", "name", ",", "value", ")", "log_mon", ".", "info", "(", "{", "'metric_name'", ":", "name", ",", "'value'...
simplest monitoring function to be aggregated with sum
[ "simplest", "monitoring", "function", "to", "be", "aggregated", "with", "sum" ]
train
https://github.com/tipsi/tipsi_tools/blob/1aba960c9890ceef2fb5e215b98b1646056ee58e/tipsi_tools/monitoring.py#L11-L16
alfredodeza/notario
notario/store.py
create_store
def create_store(): """ A helper for setting the _proxy and slapping the store object for us. :return: A thread-local storage as a dictionary """ new_storage = _proxy('store') _state.store = type('store', (object,), {}) new_storage.store = dict() return new_storage.store
python
def create_store(): """ A helper for setting the _proxy and slapping the store object for us. :return: A thread-local storage as a dictionary """ new_storage = _proxy('store') _state.store = type('store', (object,), {}) new_storage.store = dict() return new_storage.store
[ "def", "create_store", "(", ")", ":", "new_storage", "=", "_proxy", "(", "'store'", ")", "_state", ".", "store", "=", "type", "(", "'store'", ",", "(", "object", ",", ")", ",", "{", "}", ")", "new_storage", ".", "store", "=", "dict", "(", ")", "ret...
A helper for setting the _proxy and slapping the store object for us. :return: A thread-local storage as a dictionary
[ "A", "helper", "for", "setting", "the", "_proxy", "and", "slapping", "the", "store", "object", "for", "us", "." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/store.py#L25-L35
tipsi/tipsi_tools
tipsi_tools/drf/__init__.py
use_form
def use_form(form_class, request=None, **top_kwargs): """ Validate request (query_params or request body with args from url) with serializer and pass validated data dict to the view function instead of request object. """ def validated_form(request, **kwargs): # import ipdb; ipdb.set_trace(...
python
def use_form(form_class, request=None, **top_kwargs): """ Validate request (query_params or request body with args from url) with serializer and pass validated data dict to the view function instead of request object. """ def validated_form(request, **kwargs): # import ipdb; ipdb.set_trace(...
[ "def", "use_form", "(", "form_class", ",", "request", "=", "None", ",", "*", "*", "top_kwargs", ")", ":", "def", "validated_form", "(", "request", ",", "*", "*", "kwargs", ")", ":", "# import ipdb; ipdb.set_trace()", "data", "=", "request", ".", "query_param...
Validate request (query_params or request body with args from url) with serializer and pass validated data dict to the view function instead of request object.
[ "Validate", "request", "(", "query_params", "or", "request", "body", "with", "args", "from", "url", ")", "with", "serializer", "and", "pass", "validated", "data", "dict", "to", "the", "view", "function", "instead", "of", "request", "object", "." ]
train
https://github.com/tipsi/tipsi_tools/blob/1aba960c9890ceef2fb5e215b98b1646056ee58e/tipsi_tools/drf/__init__.py#L5-L49
craigahobbs/chisel
src/chisel/request.py
request
def request(request_callback=None, **kwargs): """ Chisel request decorator """ if request_callback is None: return lambda fn: request(fn, **kwargs) else: return Request(request_callback, **kwargs).decorate_module(request_callback)
python
def request(request_callback=None, **kwargs): """ Chisel request decorator """ if request_callback is None: return lambda fn: request(fn, **kwargs) else: return Request(request_callback, **kwargs).decorate_module(request_callback)
[ "def", "request", "(", "request_callback", "=", "None", ",", "*", "*", "kwargs", ")", ":", "if", "request_callback", "is", "None", ":", "return", "lambda", "fn", ":", "request", "(", "fn", ",", "*", "*", "kwargs", ")", "else", ":", "return", "Request",...
Chisel request decorator
[ "Chisel", "request", "decorator" ]
train
https://github.com/craigahobbs/chisel/blob/d306a9eae2ff757647c6ca1c933bc944efa5c326/src/chisel/request.py#L13-L21
bierschenk/ode
examples/example_2d_orbit.py
dx_orbit_sys
def dx_orbit_sys(t, X): '''X = [ m1x, m1y, m2x, m2y, m3x, m3y, m4x, m4y, m1vx, m1vy, m2vx, m2vy, m3vx, m3vy, m4vx, m4vy ] ''' (m1x, m1y, m2x, m2y, m3x, m3y, m4x, m4y, m1vx, m1vy, m2vx, m2vy, m3vx, m3vy, m4vx, m4vy) = X m_moon1 = 7.34...
python
def dx_orbit_sys(t, X): '''X = [ m1x, m1y, m2x, m2y, m3x, m3y, m4x, m4y, m1vx, m1vy, m2vx, m2vy, m3vx, m3vy, m4vx, m4vy ] ''' (m1x, m1y, m2x, m2y, m3x, m3y, m4x, m4y, m1vx, m1vy, m2vx, m2vy, m3vx, m3vy, m4vx, m4vy) = X m_moon1 = 7.34...
[ "def", "dx_orbit_sys", "(", "t", ",", "X", ")", ":", "(", "m1x", ",", "m1y", ",", "m2x", ",", "m2y", ",", "m3x", ",", "m3y", ",", "m4x", ",", "m4y", ",", "m1vx", ",", "m1vy", ",", "m2vx", ",", "m2vy", ",", "m3vx", ",", "m3vy", ",", "m4vx", ...
X = [ m1x, m1y, m2x, m2y, m3x, m3y, m4x, m4y, m1vx, m1vy, m2vx, m2vy, m3vx, m3vy, m4vx, m4vy ]
[ "X", "=", "[", "m1x", "m1y", "m2x", "m2y", "m3x", "m3y", "m4x", "m4y", "m1vx", "m1vy", "m2vx", "m2vy", "m3vx", "m3vy", "m4vx", "m4vy", "]" ]
train
https://github.com/bierschenk/ode/blob/01fb714874926f0988a4bb250d2a0c8a2429e4f0/examples/example_2d_orbit.py#L7-L76
Parsely/redis-fluster
fluster/penalty_box.py
PenaltyBox.add
def add(self, client): """Add a client to the penalty box.""" if client.pool_id in self._client_ids: log.info("%r is already in the penalty box. Ignoring.", client) return release = time.time() + self._min_wait heapq.heappush(self._clients, (release, (client, self...
python
def add(self, client): """Add a client to the penalty box.""" if client.pool_id in self._client_ids: log.info("%r is already in the penalty box. Ignoring.", client) return release = time.time() + self._min_wait heapq.heappush(self._clients, (release, (client, self...
[ "def", "add", "(", "self", ",", "client", ")", ":", "if", "client", ".", "pool_id", "in", "self", ".", "_client_ids", ":", "log", ".", "info", "(", "\"%r is already in the penalty box. Ignoring.\"", ",", "client", ")", "return", "release", "=", "time", ".", ...
Add a client to the penalty box.
[ "Add", "a", "client", "to", "the", "penalty", "box", "." ]
train
https://github.com/Parsely/redis-fluster/blob/9fb3ccdc3e0b24906520cac1e933a775e8dfbd99/fluster/penalty_box.py#L21-L28
Parsely/redis-fluster
fluster/penalty_box.py
PenaltyBox.get
def get(self): """Get any clients ready to be used. :returns: Iterable of redis clients """ now = time.time() while self._clients and self._clients[0][0] < now: _, (client, last_wait) = heapq.heappop(self._clients) connect_start = time.time() ...
python
def get(self): """Get any clients ready to be used. :returns: Iterable of redis clients """ now = time.time() while self._clients and self._clients[0][0] < now: _, (client, last_wait) = heapq.heappop(self._clients) connect_start = time.time() ...
[ "def", "get", "(", "self", ")", ":", "now", "=", "time", ".", "time", "(", ")", "while", "self", ".", "_clients", "and", "self", ".", "_clients", "[", "0", "]", "[", "0", "]", "<", "now", ":", "_", ",", "(", "client", ",", "last_wait", ")", "...
Get any clients ready to be used. :returns: Iterable of redis clients
[ "Get", "any", "clients", "ready", "to", "be", "used", "." ]
train
https://github.com/Parsely/redis-fluster/blob/9fb3ccdc3e0b24906520cac1e933a775e8dfbd99/fluster/penalty_box.py#L30-L52
alfredodeza/notario
notario/validators/types.py
string
def string(_object): """ Validates a given input is of type string. Example usage:: data = {'a' : 21} schema = (string, 21) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argum...
python
def string(_object): """ Validates a given input is of type string. Example usage:: data = {'a' : 21} schema = (string, 21) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argum...
[ "def", "string", "(", "_object", ")", ":", "if", "is_callable", "(", "_object", ")", ":", "_validator", "=", "_object", "@", "wraps", "(", "_validator", ")", "def", "decorated", "(", "value", ")", ":", "ensure", "(", "isinstance", "(", "value", ",", "b...
Validates a given input is of type string. Example usage:: data = {'a' : 21} schema = (string, 21) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argument is a callable, the decorating...
[ "Validates", "a", "given", "input", "is", "of", "type", "string", "." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L10-L34
alfredodeza/notario
notario/validators/types.py
boolean
def boolean(_object): """ Validates a given input is of type boolean. Example usage:: data = {'a' : True} schema = ('a', boolean) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the...
python
def boolean(_object): """ Validates a given input is of type boolean. Example usage:: data = {'a' : True} schema = ('a', boolean) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the...
[ "def", "boolean", "(", "_object", ")", ":", "if", "is_callable", "(", "_object", ")", ":", "_validator", "=", "_object", "@", "wraps", "(", "_validator", ")", "def", "decorated", "(", "value", ")", ":", "ensure", "(", "isinstance", "(", "value", ",", "...
Validates a given input is of type boolean. Example usage:: data = {'a' : True} schema = ('a', boolean) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argument is a callable, the decor...
[ "Validates", "a", "given", "input", "is", "of", "type", "boolean", "." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L37-L62
alfredodeza/notario
notario/validators/types.py
dictionary
def dictionary(_object, *args): """ Validates a given input is of type dictionary. Example usage:: data = {'a' : {'b': 1}} schema = ('a', dictionary) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. n...
python
def dictionary(_object, *args): """ Validates a given input is of type dictionary. Example usage:: data = {'a' : {'b': 1}} schema = ('a', dictionary) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. n...
[ "def", "dictionary", "(", "_object", ",", "*", "args", ")", ":", "error_msg", "=", "'not of type dictionary'", "if", "is_callable", "(", "_object", ")", ":", "_validator", "=", "_object", "@", "wraps", "(", "_validator", ")", "def", "decorated", "(", "value"...
Validates a given input is of type dictionary. Example usage:: data = {'a' : {'b': 1}} schema = ('a', dictionary) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argument is a callable,...
[ "Validates", "a", "given", "input", "is", "of", "type", "dictionary", "." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L66-L98
alfredodeza/notario
notario/validators/types.py
array
def array(_object): """ Validates a given input is of type list. Example usage:: data = {'a' : [1,2]} schema = ('a', array) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argum...
python
def array(_object): """ Validates a given input is of type list. Example usage:: data = {'a' : [1,2]} schema = ('a', array) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argum...
[ "def", "array", "(", "_object", ")", ":", "if", "is_callable", "(", "_object", ")", ":", "_validator", "=", "_object", "@", "wraps", "(", "_validator", ")", "def", "decorated", "(", "value", ")", ":", "ensure", "(", "isinstance", "(", "value", ",", "li...
Validates a given input is of type list. Example usage:: data = {'a' : [1,2]} schema = ('a', array) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argument is a callable, the decoratin...
[ "Validates", "a", "given", "input", "is", "of", "type", "list", "." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L101-L126
alfredodeza/notario
notario/validators/types.py
integer
def integer(_object): """ Validates a given input is of type int.. Example usage:: data = {'a' : 21} schema = ('a', integer) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argu...
python
def integer(_object): """ Validates a given input is of type int.. Example usage:: data = {'a' : 21} schema = ('a', integer) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argu...
[ "def", "integer", "(", "_object", ")", ":", "if", "is_callable", "(", "_object", ")", ":", "_validator", "=", "_object", "@", "wraps", "(", "_validator", ")", "def", "decorated", "(", "value", ")", ":", "ensure", "(", "isinstance", "(", "value", ",", "...
Validates a given input is of type int.. Example usage:: data = {'a' : 21} schema = ('a', integer) You can also use this as a decorator, as a way to check for the input before it even hits a validator you may be writing. .. note:: If the argument is a callable, the decorating...
[ "Validates", "a", "given", "input", "is", "of", "type", "int", ".." ]
train
https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L129-L153
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.constant
def constant(cls, value: Value, dtype: tf.DType = tf.float32) -> 'TensorFluent': '''Returns a constant `value` TensorFluent with given `dtype`. Args: value: The constant value. dtype: The output's data type. Returns: A constant Tensor...
python
def constant(cls, value: Value, dtype: tf.DType = tf.float32) -> 'TensorFluent': '''Returns a constant `value` TensorFluent with given `dtype`. Args: value: The constant value. dtype: The output's data type. Returns: A constant Tensor...
[ "def", "constant", "(", "cls", ",", "value", ":", "Value", ",", "dtype", ":", "tf", ".", "DType", "=", "tf", ".", "float32", ")", "->", "'TensorFluent'", ":", "t", "=", "tf", ".", "constant", "(", "value", ",", "dtype", "=", "dtype", ")", "scope", ...
Returns a constant `value` TensorFluent with given `dtype`. Args: value: The constant value. dtype: The output's data type. Returns: A constant TensorFluent.
[ "Returns", "a", "constant", "value", "TensorFluent", "with", "given", "dtype", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L67-L82
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.Bernoulli
def Bernoulli(cls, mean: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Bernoulli sampling op with given mean parameter. Args: mean: The mean parameter of the Bernoulli distribution. bat...
python
def Bernoulli(cls, mean: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Bernoulli sampling op with given mean parameter. Args: mean: The mean parameter of the Bernoulli distribution. bat...
[ "def", "Bernoulli", "(", "cls", ",", "mean", ":", "'TensorFluent'", ",", "batch_size", ":", "Optional", "[", "int", "]", "=", "None", ")", "->", "Tuple", "[", "Distribution", ",", "'TensorFluent'", "]", ":", "probs", "=", "mean", ".", "tensor", "dist", ...
Returns a TensorFluent for the Bernoulli sampling op with given mean parameter. Args: mean: The mean parameter of the Bernoulli distribution. batch_size: The size of the batch (optional). Returns: The Bernoulli distribution and a TensorFluent sample drawn from the d...
[ "Returns", "a", "TensorFluent", "for", "the", "Bernoulli", "sampling", "op", "with", "given", "mean", "parameter", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L85-L106
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.Uniform
def Uniform(cls, low: 'TensorFluent', high: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Uniform sampling op with given low and high parameters. Args: low: The low parameter of the Uniform...
python
def Uniform(cls, low: 'TensorFluent', high: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Uniform sampling op with given low and high parameters. Args: low: The low parameter of the Uniform...
[ "def", "Uniform", "(", "cls", ",", "low", ":", "'TensorFluent'", ",", "high", ":", "'TensorFluent'", ",", "batch_size", ":", "Optional", "[", "int", "]", "=", "None", ")", "->", "Tuple", "[", "Distribution", ",", "'TensorFluent'", "]", ":", "if", "low", ...
Returns a TensorFluent for the Uniform sampling op with given low and high parameters. Args: low: The low parameter of the Uniform distribution. high: The high parameter of the Uniform distribution. batch_size: The size of the batch (optional). Returns: ...
[ "Returns", "a", "TensorFluent", "for", "the", "Uniform", "sampling", "op", "with", "given", "low", "and", "high", "parameters", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L109-L135
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.Normal
def Normal(cls, mean: 'TensorFluent', variance: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Normal sampling op with given mean and variance. Args: mean: The mean parameter of the Normal d...
python
def Normal(cls, mean: 'TensorFluent', variance: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Normal sampling op with given mean and variance. Args: mean: The mean parameter of the Normal d...
[ "def", "Normal", "(", "cls", ",", "mean", ":", "'TensorFluent'", ",", "variance", ":", "'TensorFluent'", ",", "batch_size", ":", "Optional", "[", "int", "]", "=", "None", ")", "->", "Tuple", "[", "Distribution", ",", "'TensorFluent'", "]", ":", "if", "me...
Returns a TensorFluent for the Normal sampling op with given mean and variance. Args: mean: The mean parameter of the Normal distribution. variance: The variance parameter of the Normal distribution. batch_size: The size of the batch (optional). Returns: ...
[ "Returns", "a", "TensorFluent", "for", "the", "Normal", "sampling", "op", "with", "given", "mean", "and", "variance", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L138-L166
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.Gamma
def Gamma(cls, shape: 'TensorFluent', scale: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters. Args: shape: The shape parame...
python
def Gamma(cls, shape: 'TensorFluent', scale: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters. Args: shape: The shape parame...
[ "def", "Gamma", "(", "cls", ",", "shape", ":", "'TensorFluent'", ",", "scale", ":", "'TensorFluent'", ",", "batch_size", ":", "Optional", "[", "int", "]", "=", "None", ")", "->", "Tuple", "[", "Distribution", ",", "'TensorFluent'", "]", ":", "if", "shape...
Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters. Args: shape: The shape parameter of the Gamma distribution. scale: The scale parameter of the Gamma distribution. batch_size: The size of the batch (optional). Returns: ...
[ "Returns", "a", "TensorFluent", "for", "the", "Gamma", "sampling", "op", "with", "given", "shape", "and", "scale", "parameters", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L200-L229
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.Exponential
def Exponential(cls, mean: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Exponential sampling op with given mean parameter. Args: mean: The mean parameter of the Exponential distribution. ...
python
def Exponential(cls, mean: 'TensorFluent', batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']: '''Returns a TensorFluent for the Exponential sampling op with given mean parameter. Args: mean: The mean parameter of the Exponential distribution. ...
[ "def", "Exponential", "(", "cls", ",", "mean", ":", "'TensorFluent'", ",", "batch_size", ":", "Optional", "[", "int", "]", "=", "None", ")", "->", "Tuple", "[", "Distribution", ",", "'TensorFluent'", "]", ":", "rate", "=", "1", "/", "mean", ".", "tenso...
Returns a TensorFluent for the Exponential sampling op with given mean parameter. Args: mean: The mean parameter of the Exponential distribution. batch_size: The size of the batch (optional). Returns: The Exponential distribution and a TensorFluent sample drawn from...
[ "Returns", "a", "TensorFluent", "for", "the", "Exponential", "sampling", "op", "with", "given", "mean", "parameter", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L232-L253
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.stop_gradient
def stop_gradient(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a copy of the input fluent with stop_gradient at tensor level. Args: x: The input fluent. Returns: A TensorFluent that stops backpropagation of gradient computations. ''' scope = x.s...
python
def stop_gradient(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a copy of the input fluent with stop_gradient at tensor level. Args: x: The input fluent. Returns: A TensorFluent that stops backpropagation of gradient computations. ''' scope = x.s...
[ "def", "stop_gradient", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "scope", "=", "x", ".", "scope", ".", "as_list", "(", ")", "batch", "=", "x", ".", "batch", "return", "TensorFluent", "(", "tf", ".", "stop_gradient", ...
Returns a copy of the input fluent with stop_gradient at tensor level. Args: x: The input fluent. Returns: A TensorFluent that stops backpropagation of gradient computations.
[ "Returns", "a", "copy", "of", "the", "input", "fluent", "with", "stop_gradient", "at", "tensor", "level", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L256-L267
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.stop_batch_gradient
def stop_batch_gradient(cls, x: 'TensorFluent', stop_batch: tf.Tensor) -> 'TensorFluent': '''Returns a copy of the inputs fluent with stop_gradient applied at batch level. Args: x: The input fluent. stop_batch: A boolean tf.Tensor with shape=(batch_size, ...) Returns: ...
python
def stop_batch_gradient(cls, x: 'TensorFluent', stop_batch: tf.Tensor) -> 'TensorFluent': '''Returns a copy of the inputs fluent with stop_gradient applied at batch level. Args: x: The input fluent. stop_batch: A boolean tf.Tensor with shape=(batch_size, ...) Returns: ...
[ "def", "stop_batch_gradient", "(", "cls", ",", "x", ":", "'TensorFluent'", ",", "stop_batch", ":", "tf", ".", "Tensor", ")", "->", "'TensorFluent'", ":", "scope", "=", "x", ".", "scope", ".", "as_list", "(", ")", "batch", "=", "x", ".", "batch", "tenso...
Returns a copy of the inputs fluent with stop_gradient applied at batch level. Args: x: The input fluent. stop_batch: A boolean tf.Tensor with shape=(batch_size, ...) Returns: A TensorFluent that conditionally stops backpropagation of gradient computations.
[ "Returns", "a", "copy", "of", "the", "inputs", "fluent", "with", "stop_gradient", "applied", "at", "batch", "level", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L270-L283
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.abs
def abs(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the abs function. Args: x: The input fluent. Returns: A TensorFluent wrapping the abs function. ''' return cls._unary_op(x, tf.abs, tf.float32)
python
def abs(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the abs function. Args: x: The input fluent. Returns: A TensorFluent wrapping the abs function. ''' return cls._unary_op(x, tf.abs, tf.float32)
[ "def", "abs", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "abs", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the abs function. Args: x: The input fluent. Returns: A TensorFluent wrapping the abs function.
[ "Returns", "a", "TensorFluent", "for", "the", "abs", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L286-L295
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.exp
def exp(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the exp function. Args: x: The input fluent. Returns: A TensorFluent wrapping the exp function. ''' return cls._unary_op(x, tf.exp, tf.float32)
python
def exp(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the exp function. Args: x: The input fluent. Returns: A TensorFluent wrapping the exp function. ''' return cls._unary_op(x, tf.exp, tf.float32)
[ "def", "exp", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "exp", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the exp function. Args: x: The input fluent. Returns: A TensorFluent wrapping the exp function.
[ "Returns", "a", "TensorFluent", "for", "the", "exp", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L298-L307
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.log
def log(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the log function. Args: x: The input fluent. Returns: A TensorFluent wrapping the log function. ''' return cls._unary_op(x, tf.log, tf.float32)
python
def log(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the log function. Args: x: The input fluent. Returns: A TensorFluent wrapping the log function. ''' return cls._unary_op(x, tf.log, tf.float32)
[ "def", "log", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "log", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the log function. Args: x: The input fluent. Returns: A TensorFluent wrapping the log function.
[ "Returns", "a", "TensorFluent", "for", "the", "log", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L310-L319
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.sqrt
def sqrt(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the sqrt function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sqrt function. ''' return cls._unary_op(x, tf.sqrt, tf.float32)
python
def sqrt(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the sqrt function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sqrt function. ''' return cls._unary_op(x, tf.sqrt, tf.float32)
[ "def", "sqrt", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "sqrt", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the sqrt function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sqrt function.
[ "Returns", "a", "TensorFluent", "for", "the", "sqrt", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L322-L331
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.cos
def cos(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the cos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the cos function. ''' return cls._unary_op(x, tf.cos, tf.float32)
python
def cos(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the cos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the cos function. ''' return cls._unary_op(x, tf.cos, tf.float32)
[ "def", "cos", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "cos", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the cos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the cos function.
[ "Returns", "a", "TensorFluent", "for", "the", "cos", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L334-L343
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.sin
def sin(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the sin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sin function. ''' return cls._unary_op(x, tf.sin, tf.float32)
python
def sin(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the sin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sin function. ''' return cls._unary_op(x, tf.sin, tf.float32)
[ "def", "sin", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "sin", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the sin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the sin function.
[ "Returns", "a", "TensorFluent", "for", "the", "sin", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L346-L355
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.tan
def tan(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the tan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the tan function. ''' return cls._unary_op(x, tf.tan, tf.float32)
python
def tan(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the tan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the tan function. ''' return cls._unary_op(x, tf.tan, tf.float32)
[ "def", "tan", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "tan", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the tan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the tan function.
[ "Returns", "a", "TensorFluent", "for", "the", "tan", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L358-L367
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.acos
def acos(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arccos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arccos function. ''' return cls._unary_op(x, tf.acos, tf.float32)
python
def acos(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arccos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arccos function. ''' return cls._unary_op(x, tf.acos, tf.float32)
[ "def", "acos", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "acos", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the arccos function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arccos function.
[ "Returns", "a", "TensorFluent", "for", "the", "arccos", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L370-L379
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.asin
def asin(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arcsin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arcsin function. ''' return cls._unary_op(x, tf.asin, tf.float32)
python
def asin(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arcsin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arcsin function. ''' return cls._unary_op(x, tf.asin, tf.float32)
[ "def", "asin", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "asin", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the arcsin function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arcsin function.
[ "Returns", "a", "TensorFluent", "for", "the", "arcsin", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L382-L391
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.atan
def atan(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arctan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arctan function. ''' return cls._unary_op(x, tf.atan2, tf.float32)
python
def atan(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the arctan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arctan function. ''' return cls._unary_op(x, tf.atan2, tf.float32)
[ "def", "atan", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "atan2", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the arctan function. Args: x: The input fluent. Returns: A TensorFluent wrapping the arctan function.
[ "Returns", "a", "TensorFluent", "for", "the", "arctan", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L394-L403
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.round
def round(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the round function. Args: x: The input fluent. Returns: A TensorFluent wrapping the round function. ''' return cls._unary_op(x, tf.round, tf.float32)
python
def round(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the round function. Args: x: The input fluent. Returns: A TensorFluent wrapping the round function. ''' return cls._unary_op(x, tf.round, tf.float32)
[ "def", "round", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "round", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the round function. Args: x: The input fluent. Returns: A TensorFluent wrapping the round function.
[ "Returns", "a", "TensorFluent", "for", "the", "round", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L406-L415
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.ceil
def ceil(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the ceil function. Args: x: The input fluent. Returns: A TensorFluent wrapping the ceil function. ''' return cls._unary_op(x, tf.ceil, tf.float32)
python
def ceil(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the ceil function. Args: x: The input fluent. Returns: A TensorFluent wrapping the ceil function. ''' return cls._unary_op(x, tf.ceil, tf.float32)
[ "def", "ceil", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "ceil", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the ceil function. Args: x: The input fluent. Returns: A TensorFluent wrapping the ceil function.
[ "Returns", "a", "TensorFluent", "for", "the", "ceil", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L418-L427
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.floor
def floor(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the floor function. Args: x: The input fluent. Returns: A TensorFluent wrapping the floor function. ''' return cls._unary_op(x, tf.floor, tf.float32)
python
def floor(cls, x: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the floor function. Args: x: The input fluent. Returns: A TensorFluent wrapping the floor function. ''' return cls._unary_op(x, tf.floor, tf.float32)
[ "def", "floor", "(", "cls", ",", "x", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_unary_op", "(", "x", ",", "tf", ".", "floor", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the floor function. Args: x: The input fluent. Returns: A TensorFluent wrapping the floor function.
[ "Returns", "a", "TensorFluent", "for", "the", "floor", "function", "." ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L430-L439
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.pow
def pow(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the pow function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the pow function. ''' return...
python
def pow(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the pow function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the pow function. ''' return...
[ "def", "pow", "(", "cls", ",", "x", ":", "'TensorFluent'", ",", "y", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_binary_op", "(", "x", ",", "y", ",", "tf", ".", "pow", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the pow function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the pow function.
[ "Returns", "a", "TensorFluent", "for", "the", "pow", "function", ".", "TensorFluent" ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L442-L452
thiagopbueno/rddl2tf
rddl2tf/fluent.py
TensorFluent.max
def max(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the maximum function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the maximum function. ''' ...
python
def max(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent': '''Returns a TensorFluent for the maximum function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the maximum function. ''' ...
[ "def", "max", "(", "cls", ",", "x", ":", "'TensorFluent'", ",", "y", ":", "'TensorFluent'", ")", "->", "'TensorFluent'", ":", "return", "cls", ".", "_binary_op", "(", "x", ",", "y", ",", "tf", ".", "maximum", ",", "tf", ".", "float32", ")" ]
Returns a TensorFluent for the maximum function.TensorFluent Args: x: The first operand. y: The second operand. Returns: A TensorFluent wrapping the maximum function.
[ "Returns", "a", "TensorFluent", "for", "the", "maximum", "function", ".", "TensorFluent" ]
train
https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L455-L465