repository_name stringlengths 7 55 | func_path_in_repository stringlengths 4 223 | func_name stringlengths 1 134 | whole_func_string stringlengths 75 104k | language stringclasses 1
value | func_code_string stringlengths 75 104k | func_code_tokens listlengths 19 28.4k | func_documentation_string stringlengths 1 46.9k | func_documentation_tokens listlengths 1 1.97k | split_name stringclasses 1
value | func_code_url stringlengths 87 315 |
|---|---|---|---|---|---|---|---|---|---|---|
sashs/filebytes | filebytes/pe.py | PE._getSectionForDataDirectoryEntry | def _getSectionForDataDirectoryEntry(self, data_directory_entry, sections):
"""Returns the section which contains the data of DataDirectory"""
for section in sections:
if data_directory_entry.VirtualAddress >= section.header.VirtualAddress and \
data_directory_entry.VirtualAddres... | python | def _getSectionForDataDirectoryEntry(self, data_directory_entry, sections):
"""Returns the section which contains the data of DataDirectory"""
for section in sections:
if data_directory_entry.VirtualAddress >= section.header.VirtualAddress and \
data_directory_entry.VirtualAddres... | [
"def",
"_getSectionForDataDirectoryEntry",
"(",
"self",
",",
"data_directory_entry",
",",
"sections",
")",
":",
"for",
"section",
"in",
"sections",
":",
"if",
"data_directory_entry",
".",
"VirtualAddress",
">=",
"section",
".",
"header",
".",
"VirtualAddress",
"and"... | Returns the section which contains the data of DataDirectory | [
"Returns",
"the",
"section",
"which",
"contains",
"the",
"data",
"of",
"DataDirectory"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L541-L547 |
sashs/filebytes | filebytes/pe.py | PE._parseDataDirectory | def _parseDataDirectory(self, data, sections, imageNtHeaders):
"""Parses the entries of the DataDirectory and returns a list of the content"""
data_directory_data_list = [None for i in range(15)]
# parse DataDirectory[Export]
export_data_directory = imageNtHeaders.header.OptionalHeader.... | python | def _parseDataDirectory(self, data, sections, imageNtHeaders):
"""Parses the entries of the DataDirectory and returns a list of the content"""
data_directory_data_list = [None for i in range(15)]
# parse DataDirectory[Export]
export_data_directory = imageNtHeaders.header.OptionalHeader.... | [
"def",
"_parseDataDirectory",
"(",
"self",
",",
"data",
",",
"sections",
",",
"imageNtHeaders",
")",
":",
"data_directory_data_list",
"=",
"[",
"None",
"for",
"i",
"in",
"range",
"(",
"15",
")",
"]",
"# parse DataDirectory[Export]",
"export_data_directory",
"=",
... | Parses the entries of the DataDirectory and returns a list of the content | [
"Parses",
"the",
"entries",
"of",
"the",
"DataDirectory",
"and",
"returns",
"a",
"list",
"of",
"the",
"content"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L549-L571 |
sashs/filebytes | filebytes/pe.py | PE._parseDataDirectoryExport | def _parseDataDirectoryExport(self, data, dataDirectoryEntry, exportSection):
"""Parses the EmportDataDirectory and returns an instance of ExportDirectoryData"""
if not exportSection:
return
functions = []
export_directory = IMAGE_EXPORT_DIRECTORY.from_buffer(exportSection.ra... | python | def _parseDataDirectoryExport(self, data, dataDirectoryEntry, exportSection):
"""Parses the EmportDataDirectory and returns an instance of ExportDirectoryData"""
if not exportSection:
return
functions = []
export_directory = IMAGE_EXPORT_DIRECTORY.from_buffer(exportSection.ra... | [
"def",
"_parseDataDirectoryExport",
"(",
"self",
",",
"data",
",",
"dataDirectoryEntry",
",",
"exportSection",
")",
":",
"if",
"not",
"exportSection",
":",
"return",
"functions",
"=",
"[",
"]",
"export_directory",
"=",
"IMAGE_EXPORT_DIRECTORY",
".",
"from_buffer",
... | Parses the EmportDataDirectory and returns an instance of ExportDirectoryData | [
"Parses",
"the",
"EmportDataDirectory",
"and",
"returns",
"an",
"instance",
"of",
"ExportDirectoryData"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L573-L601 |
sashs/filebytes | filebytes/pe.py | PE._parseDataDirectoryImport | def _parseDataDirectoryImport(self, dataDirectoryEntry, importSection):
"""Parses the ImportDataDirectory and returns a list of ImportDescriptorData"""
if not importSection:
return
raw_bytes = (c_ubyte * dataDirectoryEntry.Size).from_buffer(importSection.raw, to_offset(dataDirector... | python | def _parseDataDirectoryImport(self, dataDirectoryEntry, importSection):
"""Parses the ImportDataDirectory and returns a list of ImportDescriptorData"""
if not importSection:
return
raw_bytes = (c_ubyte * dataDirectoryEntry.Size).from_buffer(importSection.raw, to_offset(dataDirector... | [
"def",
"_parseDataDirectoryImport",
"(",
"self",
",",
"dataDirectoryEntry",
",",
"importSection",
")",
":",
"if",
"not",
"importSection",
":",
"return",
"raw_bytes",
"=",
"(",
"c_ubyte",
"*",
"dataDirectoryEntry",
".",
"Size",
")",
".",
"from_buffer",
"(",
"impo... | Parses the ImportDataDirectory and returns a list of ImportDescriptorData | [
"Parses",
"the",
"ImportDataDirectory",
"and",
"returns",
"a",
"list",
"of",
"ImportDescriptorData"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L603-L629 |
sashs/filebytes | filebytes/pe.py | PE.__parseThunks | def __parseThunks(self, thunkRVA, importSection):
"""Parses the thunks and returns a list"""
offset = to_offset(thunkRVA, importSection)
table_offset = 0
thunks = []
while True:
thunk = IMAGE_THUNK_DATA.from_buffer(importSection.raw, offset)
offset += size... | python | def __parseThunks(self, thunkRVA, importSection):
"""Parses the thunks and returns a list"""
offset = to_offset(thunkRVA, importSection)
table_offset = 0
thunks = []
while True:
thunk = IMAGE_THUNK_DATA.from_buffer(importSection.raw, offset)
offset += size... | [
"def",
"__parseThunks",
"(",
"self",
",",
"thunkRVA",
",",
"importSection",
")",
":",
"offset",
"=",
"to_offset",
"(",
"thunkRVA",
",",
"importSection",
")",
"table_offset",
"=",
"0",
"thunks",
"=",
"[",
"]",
"while",
"True",
":",
"thunk",
"=",
"IMAGE_THUN... | Parses the thunks and returns a list | [
"Parses",
"the",
"thunks",
"and",
"returns",
"a",
"list"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L672-L687 |
sashs/filebytes | filebytes/pe.py | PE.__parseThunkData | def __parseThunkData(self, thunk,importSection):
"""Parses the data of a thunk and sets the data"""
offset = to_offset(thunk.header.AddressOfData, importSection)
if 0xf0000000 & thunk.header.AddressOfData == 0x80000000:
thunk.ordinal = thunk.header.AddressOfData & 0x0fffffff
... | python | def __parseThunkData(self, thunk,importSection):
"""Parses the data of a thunk and sets the data"""
offset = to_offset(thunk.header.AddressOfData, importSection)
if 0xf0000000 & thunk.header.AddressOfData == 0x80000000:
thunk.ordinal = thunk.header.AddressOfData & 0x0fffffff
... | [
"def",
"__parseThunkData",
"(",
"self",
",",
"thunk",
",",
"importSection",
")",
":",
"offset",
"=",
"to_offset",
"(",
"thunk",
".",
"header",
".",
"AddressOfData",
",",
"importSection",
")",
"if",
"0xf0000000",
"&",
"thunk",
".",
"header",
".",
"AddressOfDa... | Parses the data of a thunk and sets the data | [
"Parses",
"the",
"data",
"of",
"a",
"thunk",
"and",
"sets",
"the",
"data"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/pe.py#L689-L699 |
sashs/filebytes | filebytes/ctypes_helper.py | get_ptr | def get_ptr(data, offset=None, ptr_type=ctypes.c_void_p):
"""Returns a void pointer to the data"""
ptr = ctypes.cast(ctypes.pointer(data), ctypes.c_void_p)
if offset:
ptr = ctypes.c_void_p(ptr.value + offset)
if ptr_type != ctypes.c_void_p:
ptr = ctypes.cast(ptr, ptr_type)
return ... | python | def get_ptr(data, offset=None, ptr_type=ctypes.c_void_p):
"""Returns a void pointer to the data"""
ptr = ctypes.cast(ctypes.pointer(data), ctypes.c_void_p)
if offset:
ptr = ctypes.c_void_p(ptr.value + offset)
if ptr_type != ctypes.c_void_p:
ptr = ctypes.cast(ptr, ptr_type)
return ... | [
"def",
"get_ptr",
"(",
"data",
",",
"offset",
"=",
"None",
",",
"ptr_type",
"=",
"ctypes",
".",
"c_void_p",
")",
":",
"ptr",
"=",
"ctypes",
".",
"cast",
"(",
"ctypes",
".",
"pointer",
"(",
"data",
")",
",",
"ctypes",
".",
"c_void_p",
")",
"if",
"of... | Returns a void pointer to the data | [
"Returns",
"a",
"void",
"pointer",
"to",
"the",
"data"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/ctypes_helper.py#L33-L43 |
sashs/filebytes | filebytes/ctypes_helper.py | to_ubyte_array | def to_ubyte_array(barray):
"""Returns a c_ubyte_array filled with the given data of a bytearray or bytes"""
bs = (ctypes.c_ubyte * len(barray))()
pack_into('%ds' % len(barray), bs, 0, barray)
return bs | python | def to_ubyte_array(barray):
"""Returns a c_ubyte_array filled with the given data of a bytearray or bytes"""
bs = (ctypes.c_ubyte * len(barray))()
pack_into('%ds' % len(barray), bs, 0, barray)
return bs | [
"def",
"to_ubyte_array",
"(",
"barray",
")",
":",
"bs",
"=",
"(",
"ctypes",
".",
"c_ubyte",
"*",
"len",
"(",
"barray",
")",
")",
"(",
")",
"pack_into",
"(",
"'%ds'",
"%",
"len",
"(",
"barray",
")",
",",
"bs",
",",
"0",
",",
"barray",
")",
"return... | Returns a c_ubyte_array filled with the given data of a bytearray or bytes | [
"Returns",
"a",
"c_ubyte_array",
"filled",
"with",
"the",
"given",
"data",
"of",
"a",
"bytearray",
"or",
"bytes"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/ctypes_helper.py#L48-L53 |
sashs/filebytes | filebytes/binary.py | Binary._readFile | def _readFile(self, fileName):
"""
Returns the bytes of the file.
"""
with open(fileName, 'rb') as binFile:
b = binFile.read()
return to_ubyte_array(b) | python | def _readFile(self, fileName):
"""
Returns the bytes of the file.
"""
with open(fileName, 'rb') as binFile:
b = binFile.read()
return to_ubyte_array(b) | [
"def",
"_readFile",
"(",
"self",
",",
"fileName",
")",
":",
"with",
"open",
"(",
"fileName",
",",
"'rb'",
")",
"as",
"binFile",
":",
"b",
"=",
"binFile",
".",
"read",
"(",
")",
"return",
"to_ubyte_array",
"(",
"b",
")"
] | Returns the bytes of the file. | [
"Returns",
"the",
"bytes",
"of",
"the",
"file",
"."
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/binary.py#L71-L77 |
sashs/filebytes | filebytes/elf.py | ELF._getSuitableClasses | def _getSuitableClasses(self, data):
"""Returns the class which holds the suitable classes for the loaded file"""
classes = None
if data[EI.CLASS] == ELFCLASS.BITS_32:
if data[EI.DATA] == ELFDATA.LSB:
classes = LSB_32
elif data[EI.DATA] == ELFDATA.MSB:
... | python | def _getSuitableClasses(self, data):
"""Returns the class which holds the suitable classes for the loaded file"""
classes = None
if data[EI.CLASS] == ELFCLASS.BITS_32:
if data[EI.DATA] == ELFDATA.LSB:
classes = LSB_32
elif data[EI.DATA] == ELFDATA.MSB:
... | [
"def",
"_getSuitableClasses",
"(",
"self",
",",
"data",
")",
":",
"classes",
"=",
"None",
"if",
"data",
"[",
"EI",
".",
"CLASS",
"]",
"==",
"ELFCLASS",
".",
"BITS_32",
":",
"if",
"data",
"[",
"EI",
".",
"DATA",
"]",
"==",
"ELFDATA",
".",
"LSB",
":"... | Returns the class which holds the suitable classes for the loaded file | [
"Returns",
"the",
"class",
"which",
"holds",
"the",
"suitable",
"classes",
"for",
"the",
"loaded",
"file"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L881-L896 |
sashs/filebytes | filebytes/elf.py | ELF._parseElfHeader | def _parseElfHeader(self, data):
"""Returns the elf header"""
ehdr = self.__classes.EHDR.from_buffer(data)
return EhdrData(header=ehdr) | python | def _parseElfHeader(self, data):
"""Returns the elf header"""
ehdr = self.__classes.EHDR.from_buffer(data)
return EhdrData(header=ehdr) | [
"def",
"_parseElfHeader",
"(",
"self",
",",
"data",
")",
":",
"ehdr",
"=",
"self",
".",
"__classes",
".",
"EHDR",
".",
"from_buffer",
"(",
"data",
")",
"return",
"EhdrData",
"(",
"header",
"=",
"ehdr",
")"
] | Returns the elf header | [
"Returns",
"the",
"elf",
"header"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L898-L901 |
sashs/filebytes | filebytes/elf.py | ELF._parseSegments | def _parseSegments(self, data, elfHeader):
"""Return a list of segments"""
offset = elfHeader.header.e_phoff
segments = []
for i in range(elfHeader.header.e_phnum):
phdr = self.__classes.PHDR.from_buffer(data, offset)
segment_bytes = (c_ubyte * phdr.p_filesz).from... | python | def _parseSegments(self, data, elfHeader):
"""Return a list of segments"""
offset = elfHeader.header.e_phoff
segments = []
for i in range(elfHeader.header.e_phnum):
phdr = self.__classes.PHDR.from_buffer(data, offset)
segment_bytes = (c_ubyte * phdr.p_filesz).from... | [
"def",
"_parseSegments",
"(",
"self",
",",
"data",
",",
"elfHeader",
")",
":",
"offset",
"=",
"elfHeader",
".",
"header",
".",
"e_phoff",
"segments",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"elfHeader",
".",
"header",
".",
"e_phnum",
")",
":",
... | Return a list of segments | [
"Return",
"a",
"list",
"of",
"segments"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L903-L916 |
sashs/filebytes | filebytes/elf.py | ELF._parseSections | def _parseSections(self, data, elfHeader):
"""Returns a list of sections"""
offset = elfHeader.header.e_shoff
shdrs = []
for i in range(elfHeader.header.e_shnum):
shdr = self.__classes.SHDR.from_buffer(data, offset)
section_bytes = None
ba_section_byte... | python | def _parseSections(self, data, elfHeader):
"""Returns a list of sections"""
offset = elfHeader.header.e_shoff
shdrs = []
for i in range(elfHeader.header.e_shnum):
shdr = self.__classes.SHDR.from_buffer(data, offset)
section_bytes = None
ba_section_byte... | [
"def",
"_parseSections",
"(",
"self",
",",
"data",
",",
"elfHeader",
")",
":",
"offset",
"=",
"elfHeader",
".",
"header",
".",
"e_shoff",
"shdrs",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"elfHeader",
".",
"header",
".",
"e_shnum",
")",
":",
"s... | Returns a list of sections | [
"Returns",
"a",
"list",
"of",
"sections"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L918-L939 |
sashs/filebytes | filebytes/elf.py | ELF._parseSymbols | def _parseSymbols(self, sections):
"""Sets a list of symbols in each DYNSYM and SYMTAB section"""
for section in sections:
strtab = sections[section.header.sh_link]
if section.header.sh_type in (int(SHT.DYNSYM), int(SHT.SYMTAB)):
section.symbols = self.__parseSymb... | python | def _parseSymbols(self, sections):
"""Sets a list of symbols in each DYNSYM and SYMTAB section"""
for section in sections:
strtab = sections[section.header.sh_link]
if section.header.sh_type in (int(SHT.DYNSYM), int(SHT.SYMTAB)):
section.symbols = self.__parseSymb... | [
"def",
"_parseSymbols",
"(",
"self",
",",
"sections",
")",
":",
"for",
"section",
"in",
"sections",
":",
"strtab",
"=",
"sections",
"[",
"section",
".",
"header",
".",
"sh_link",
"]",
"if",
"section",
".",
"header",
".",
"sh_type",
"in",
"(",
"int",
"(... | Sets a list of symbols in each DYNSYM and SYMTAB section | [
"Sets",
"a",
"list",
"of",
"symbols",
"in",
"each",
"DYNSYM",
"and",
"SYMTAB",
"section"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L941-L946 |
sashs/filebytes | filebytes/elf.py | ELF._parseRelocations | def _parseRelocations(self, sections):
"""Parses the relocations and add those to the section"""
for section in sections:
if section.header.sh_link != SHN.UNDEF and section.header.sh_type in (SHT.REL, SHT.RELA):
symbols = sections[section.header.sh_link].symbols
... | python | def _parseRelocations(self, sections):
"""Parses the relocations and add those to the section"""
for section in sections:
if section.header.sh_link != SHN.UNDEF and section.header.sh_type in (SHT.REL, SHT.RELA):
symbols = sections[section.header.sh_link].symbols
... | [
"def",
"_parseRelocations",
"(",
"self",
",",
"sections",
")",
":",
"for",
"section",
"in",
"sections",
":",
"if",
"section",
".",
"header",
".",
"sh_link",
"!=",
"SHN",
".",
"UNDEF",
"and",
"section",
".",
"header",
".",
"sh_type",
"in",
"(",
"SHT",
"... | Parses the relocations and add those to the section | [
"Parses",
"the",
"relocations",
"and",
"add",
"those",
"to",
"the",
"section"
] | train | https://github.com/sashs/filebytes/blob/41ee009832aba19603f33d1fd3483b84d6684ebf/filebytes/elf.py#L965-L971 |
pyqg/pyqg | pyqg/model.py | run_with_snapshots | def run_with_snapshots(self, tsnapstart=0., tsnapint=432000.):
"""Run the model forward, yielding to user code at specified intervals.
Parameters
----------
tsnapstart : int
The timestep at which to begin yielding.
tstapint : int
The interval at which to... | python | def run_with_snapshots(self, tsnapstart=0., tsnapint=432000.):
"""Run the model forward, yielding to user code at specified intervals.
Parameters
----------
tsnapstart : int
The timestep at which to begin yielding.
tstapint : int
The interval at which to... | [
"def",
"run_with_snapshots",
"(",
"self",
",",
"tsnapstart",
"=",
"0.",
",",
"tsnapint",
"=",
"432000.",
")",
":",
"tsnapints",
"=",
"np",
".",
"ceil",
"(",
"tsnapint",
"/",
"self",
".",
"dt",
")",
"while",
"(",
"self",
".",
"t",
"<",
"self",
".",
... | Run the model forward, yielding to user code at specified intervals.
Parameters
----------
tsnapstart : int
The timestep at which to begin yielding.
tstapint : int
The interval at which to yield. | [
"Run",
"the",
"model",
"forward",
"yielding",
"to",
"user",
"code",
"at",
"specified",
"intervals",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L210-L228 |
pyqg/pyqg | pyqg/model.py | vertical_modes | def vertical_modes(self):
""" Calculate standard vertical modes. Simply
the eigenvectors of the stretching matrix S """
evals,evecs = np.linalg.eig(-self.S)
asort = evals.argsort()
# deformation wavenumbers and radii
self.kdi2 = evals[asort]
self.radii = np... | python | def vertical_modes(self):
""" Calculate standard vertical modes. Simply
the eigenvectors of the stretching matrix S """
evals,evecs = np.linalg.eig(-self.S)
asort = evals.argsort()
# deformation wavenumbers and radii
self.kdi2 = evals[asort]
self.radii = np... | [
"def",
"vertical_modes",
"(",
"self",
")",
":",
"evals",
",",
"evecs",
"=",
"np",
".",
"linalg",
".",
"eig",
"(",
"-",
"self",
".",
"S",
")",
"asort",
"=",
"evals",
".",
"argsort",
"(",
")",
"# deformation wavenumbers and radii",
"self",
".",
"kdi2",
"... | Calculate standard vertical modes. Simply
the eigenvectors of the stretching matrix S | [
"Calculate",
"standard",
"vertical",
"modes",
".",
"Simply",
"the",
"eigenvectors",
"of",
"the",
"stretching",
"matrix",
"S"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L236-L255 |
pyqg/pyqg | pyqg/model.py | modal_projection | def modal_projection(self,p,forward=True):
""" Performs a field p into modal amplitudes pn
using the basis [pmodes]. The inverse
transform calculates p from pn"""
if forward:
pt = np.linalg.solve(self.pmodes[np.newaxis,np.newaxis],p.T).T
else:
... | python | def modal_projection(self,p,forward=True):
""" Performs a field p into modal amplitudes pn
using the basis [pmodes]. The inverse
transform calculates p from pn"""
if forward:
pt = np.linalg.solve(self.pmodes[np.newaxis,np.newaxis],p.T).T
else:
... | [
"def",
"modal_projection",
"(",
"self",
",",
"p",
",",
"forward",
"=",
"True",
")",
":",
"if",
"forward",
":",
"pt",
"=",
"np",
".",
"linalg",
".",
"solve",
"(",
"self",
".",
"pmodes",
"[",
"np",
".",
"newaxis",
",",
"np",
".",
"newaxis",
"]",
",... | Performs a field p into modal amplitudes pn
using the basis [pmodes]. The inverse
transform calculates p from pn | [
"Performs",
"a",
"field",
"p",
"into",
"modal",
"amplitudes",
"pn",
"using",
"the",
"basis",
"[",
"pmodes",
"]",
".",
"The",
"inverse",
"transform",
"calculates",
"p",
"from",
"pn"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/model.py#L257-L267 |
pyqg/pyqg | pyqg/sqg_model.py | SQGModel._initialize_background | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
# background vel.
if len(np.shape(self.U)) == 0:
self.U = (self.U * np.ones((self.ny)))
print(np.shape(self.U))
self.set_U(self.U)
# the meridional PV gradients in e... | python | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
# background vel.
if len(np.shape(self.U)) == 0:
self.U = (self.U * np.ones((self.ny)))
print(np.shape(self.U))
self.set_U(self.U)
# the meridional PV gradients in e... | [
"def",
"_initialize_background",
"(",
"self",
")",
":",
"# background vel.",
"if",
"len",
"(",
"np",
".",
"shape",
"(",
"self",
".",
"U",
")",
")",
"==",
"0",
":",
"self",
".",
"U",
"=",
"(",
"self",
".",
"U",
"*",
"np",
".",
"ones",
"(",
"(",
... | Set up background state (zonal flow and PV gradients). | [
"Set",
"up",
"background",
"state",
"(",
"zonal",
"flow",
"and",
"PV",
"gradients",
")",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L51-L67 |
pyqg/pyqg | pyqg/sqg_model.py | SQGModel._initialize_inversion_matrix | def _initialize_inversion_matrix(self):
""" the inversion """
# The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph
self.a = np.asarray(self.Nb*np.sqrt(self.wv2i))[np.newaxis, np.newaxis, :, :] | python | def _initialize_inversion_matrix(self):
""" the inversion """
# The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph
self.a = np.asarray(self.Nb*np.sqrt(self.wv2i))[np.newaxis, np.newaxis, :, :] | [
"def",
"_initialize_inversion_matrix",
"(",
"self",
")",
":",
"# The sqg model is diagonal. The inversion is simply qh = -kappa**2 ph",
"self",
".",
"a",
"=",
"np",
".",
"asarray",
"(",
"self",
".",
"Nb",
"*",
"np",
".",
"sqrt",
"(",
"self",
".",
"wv2i",
")",
")... | the inversion | [
"the",
"inversion"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L69-L72 |
pyqg/pyqg | pyqg/sqg_model.py | SQGModel.set_U | def set_U(self, U):
"""Set background zonal flow"""
self.Ubg = np.asarray(U)[np.newaxis,...] | python | def set_U(self, U):
"""Set background zonal flow"""
self.Ubg = np.asarray(U)[np.newaxis,...] | [
"def",
"set_U",
"(",
"self",
",",
"U",
")",
":",
"self",
".",
"Ubg",
"=",
"np",
".",
"asarray",
"(",
"U",
")",
"[",
"np",
".",
"newaxis",
",",
"...",
"]"
] | Set background zonal flow | [
"Set",
"background",
"zonal",
"flow"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/sqg_model.py#L77-L79 |
pyqg/pyqg | pyqg/particles.py | LagrangianParticleArray2D.step_forward_with_function | def step_forward_with_function(self, uv0fun, uv1fun, dt):
"""Advance particles using a function to determine u and v.
Parameters
----------
uv0fun : function
Called like ``uv0fun(x,y)``. Should return the velocity field
u, v at time t.
uv1fun(x,y)... | python | def step_forward_with_function(self, uv0fun, uv1fun, dt):
"""Advance particles using a function to determine u and v.
Parameters
----------
uv0fun : function
Called like ``uv0fun(x,y)``. Should return the velocity field
u, v at time t.
uv1fun(x,y)... | [
"def",
"step_forward_with_function",
"(",
"self",
",",
"uv0fun",
",",
"uv1fun",
",",
"dt",
")",
":",
"dx",
",",
"dy",
"=",
"self",
".",
"_rk4_integrate",
"(",
"self",
".",
"x",
",",
"self",
".",
"y",
",",
"uv0fun",
",",
"uv1fun",
",",
"dt",
")",
"s... | Advance particles using a function to determine u and v.
Parameters
----------
uv0fun : function
Called like ``uv0fun(x,y)``. Should return the velocity field
u, v at time t.
uv1fun(x,y) : function
Called like ``uv1fun(x,y)``. Should return th... | [
"Advance",
"particles",
"using",
"a",
"function",
"to",
"determine",
"u",
"and",
"v",
".",
"Parameters",
"----------",
"uv0fun",
":",
"function",
"Called",
"like",
"uv0fun",
"(",
"x",
"y",
")",
".",
"Should",
"return",
"the",
"velocity",
"field",
"u",
"v",... | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L65-L81 |
pyqg/pyqg | pyqg/particles.py | LagrangianParticleArray2D._rk4_integrate | def _rk4_integrate(self, x, y, uv0fun, uv1fun, dt):
"""Integrates positions x, y using velocity functions
uv0fun, uv1fun. Returns dx and dy, the displacements."""
u0, v0 = uv0fun(x, y)
k1u = dt*u0
k1v = dt*v0
x11 = self._wrap_x(x + 0.5*k1u)
y11 = self._wrap_y(y... | python | def _rk4_integrate(self, x, y, uv0fun, uv1fun, dt):
"""Integrates positions x, y using velocity functions
uv0fun, uv1fun. Returns dx and dy, the displacements."""
u0, v0 = uv0fun(x, y)
k1u = dt*u0
k1v = dt*v0
x11 = self._wrap_x(x + 0.5*k1u)
y11 = self._wrap_y(y... | [
"def",
"_rk4_integrate",
"(",
"self",
",",
"x",
",",
"y",
",",
"uv0fun",
",",
"uv1fun",
",",
"dt",
")",
":",
"u0",
",",
"v0",
"=",
"uv0fun",
"(",
"x",
",",
"y",
")",
"k1u",
"=",
"dt",
"*",
"u0",
"k1v",
"=",
"dt",
"*",
"v0",
"x11",
"=",
"sel... | Integrates positions x, y using velocity functions
uv0fun, uv1fun. Returns dx and dy, the displacements. | [
"Integrates",
"positions",
"x",
"y",
"using",
"velocity",
"functions",
"uv0fun",
"uv1fun",
".",
"Returns",
"dx",
"and",
"dy",
"the",
"displacements",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L83-L108 |
pyqg/pyqg | pyqg/particles.py | LagrangianParticleArray2D._distance | def _distance(self, x0, y0, x1, y1):
"""Utitlity function to compute distance between points."""
dx = x1-x0
dy = y1-y0
# roll displacements across the borders
if self.pix:
dx[ dx > self.Lx/2 ] -= self.Lx
dx[ dx < -self.Lx/2 ] += self.Lx
if self.piy... | python | def _distance(self, x0, y0, x1, y1):
"""Utitlity function to compute distance between points."""
dx = x1-x0
dy = y1-y0
# roll displacements across the borders
if self.pix:
dx[ dx > self.Lx/2 ] -= self.Lx
dx[ dx < -self.Lx/2 ] += self.Lx
if self.piy... | [
"def",
"_distance",
"(",
"self",
",",
"x0",
",",
"y0",
",",
"x1",
",",
"y1",
")",
":",
"dx",
"=",
"x1",
"-",
"x0",
"dy",
"=",
"y1",
"-",
"y0",
"# roll displacements across the borders",
"if",
"self",
".",
"pix",
":",
"dx",
"[",
"dx",
">",
"self",
... | Utitlity function to compute distance between points. | [
"Utitlity",
"function",
"to",
"compute",
"distance",
"between",
"points",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L124-L135 |
pyqg/pyqg | pyqg/particles.py | GriddedLagrangianParticleArray2D.interpolate_gridded_scalar | def interpolate_gridded_scalar(self, x, y, c, order=1, pad=1, offset=0):
"""Interpolate gridded scalar C to points x,y.
Parameters
----------
x, y : array-like
Points at which to interpolate
c : array-like
The scalar, assumed to be defined on the ... | python | def interpolate_gridded_scalar(self, x, y, c, order=1, pad=1, offset=0):
"""Interpolate gridded scalar C to points x,y.
Parameters
----------
x, y : array-like
Points at which to interpolate
c : array-like
The scalar, assumed to be defined on the ... | [
"def",
"interpolate_gridded_scalar",
"(",
"self",
",",
"x",
",",
"y",
",",
"c",
",",
"order",
"=",
"1",
",",
"pad",
"=",
"1",
",",
"offset",
"=",
"0",
")",
":",
"## no longer necessary because we accept pre-padded arrays",
"# assert c.shape == (self.Ny, self.Nx), 'S... | Interpolate gridded scalar C to points x,y.
Parameters
----------
x, y : array-like
Points at which to interpolate
c : array-like
The scalar, assumed to be defined on the grid.
order : int
Order of interpolation
pad : int
... | [
"Interpolate",
"gridded",
"scalar",
"C",
"to",
"points",
"x",
"y",
".",
"Parameters",
"----------",
"x",
"y",
":",
"array",
"-",
"like",
"Points",
"at",
"which",
"to",
"interpolate",
"c",
":",
"array",
"-",
"like",
"The",
"scalar",
"assumed",
"to",
"be",... | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L180-L218 |
pyqg/pyqg | pyqg/particles.py | GriddedLagrangianParticleArray2D.step_forward_with_gridded_uv | def step_forward_with_gridded_uv(self, U0, V0, U1, V1, dt, order=1):
"""Advance particles using a gridded velocity field. Because of the
Runga-Kutta timestepping, we need two velocity fields at different
times.
Parameters
----------
U0, V0 : array-like
... | python | def step_forward_with_gridded_uv(self, U0, V0, U1, V1, dt, order=1):
"""Advance particles using a gridded velocity field. Because of the
Runga-Kutta timestepping, we need two velocity fields at different
times.
Parameters
----------
U0, V0 : array-like
... | [
"def",
"step_forward_with_gridded_uv",
"(",
"self",
",",
"U0",
",",
"V0",
",",
"U1",
",",
"V1",
",",
"dt",
",",
"order",
"=",
"1",
")",
":",
"# create interpolation functions which return u and v",
"# pre-pad arrays so it only has to be done once",
"# for linear interpola... | Advance particles using a gridded velocity field. Because of the
Runga-Kutta timestepping, we need two velocity fields at different
times.
Parameters
----------
U0, V0 : array-like
Gridded velocity fields at time t - dt.
U1, V1 : array-like
... | [
"Advance",
"particles",
"using",
"a",
"gridded",
"velocity",
"field",
".",
"Because",
"of",
"the",
"Runga",
"-",
"Kutta",
"timestepping",
"we",
"need",
"two",
"velocity",
"fields",
"at",
"different",
"times",
".",
"Parameters",
"----------",
"U0",
"V0",
":",
... | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/particles.py#L223-L258 |
pyqg/pyqg | pyqg/diagnostic_tools.py | spec_var | def spec_var(model, ph):
"""Compute variance of ``p`` from Fourier coefficients ``ph``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
var_dens :... | python | def spec_var(model, ph):
"""Compute variance of ``p`` from Fourier coefficients ``ph``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
var_dens :... | [
"def",
"spec_var",
"(",
"model",
",",
"ph",
")",
":",
"var_dens",
"=",
"2.",
"*",
"np",
".",
"abs",
"(",
"ph",
")",
"**",
"2",
"/",
"model",
".",
"M",
"**",
"2",
"# only half of coefs [0] and [nx/2+1] due to symmetry in real fft2",
"var_dens",
"[",
"...",
... | Compute variance of ``p`` from Fourier coefficients ``ph``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
var_dens : float
The variance of `... | [
"Compute",
"variance",
"of",
"p",
"from",
"Fourier",
"coefficients",
"ph",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L7-L27 |
pyqg/pyqg | pyqg/diagnostic_tools.py | spec_sum | def spec_sum(ph2):
"""Compute total spectral sum of the real spectral quantity``ph^2``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph2 : real array
The field on which to compute the sum
Returns
-------
var_dens : float
... | python | def spec_sum(ph2):
"""Compute total spectral sum of the real spectral quantity``ph^2``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph2 : real array
The field on which to compute the sum
Returns
-------
var_dens : float
... | [
"def",
"spec_sum",
"(",
"ph2",
")",
":",
"ph2",
"=",
"2.",
"*",
"ph2",
"ph2",
"[",
"...",
",",
"0",
"]",
"=",
"ph2",
"[",
"...",
",",
"0",
"]",
"/",
"2.",
"ph2",
"[",
"...",
",",
"-",
"1",
"]",
"=",
"ph2",
"[",
"...",
",",
"-",
"1",
"]"... | Compute total spectral sum of the real spectral quantity``ph^2``.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph2 : real array
The field on which to compute the sum
Returns
-------
var_dens : float
The sum of `ph2` | [
"Compute",
"total",
"spectral",
"sum",
"of",
"the",
"real",
"spectral",
"quantity",
"ph^2",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L30-L50 |
pyqg/pyqg | pyqg/diagnostic_tools.py | calc_ispec | def calc_ispec(model, ph):
"""Compute isotropic spectrum `phr` of `ph` from 2D spectrum.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
kr : arra... | python | def calc_ispec(model, ph):
"""Compute isotropic spectrum `phr` of `ph` from 2D spectrum.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
kr : arra... | [
"def",
"calc_ispec",
"(",
"model",
",",
"ph",
")",
":",
"if",
"model",
".",
"kk",
".",
"max",
"(",
")",
">",
"model",
".",
"ll",
".",
"max",
"(",
")",
":",
"kmax",
"=",
"model",
".",
"ll",
".",
"max",
"(",
")",
"else",
":",
"kmax",
"=",
"mo... | Compute isotropic spectrum `phr` of `ph` from 2D spectrum.
Parameters
----------
model : pyqg.Model instance
The model object from which `ph` originates
ph : complex array
The field on which to compute the variance
Returns
-------
kr : array
isotropic wavenumber
... | [
"Compute",
"isotropic",
"spectrum",
"phr",
"of",
"ph",
"from",
"2D",
"spectrum",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/diagnostic_tools.py#L53-L86 |
pyqg/pyqg | pyqg/layered_model.py | LayeredModel._initialize_stretching_matrix | def _initialize_stretching_matrix(self):
""" Set up the stretching matrix """
self.S = np.zeros((self.nz, self.nz))
if (self.nz==2) and (self.rd) and (self.delta):
self.del1 = self.delta/(self.delta+1.)
self.del2 = (self.delta+1.)**-1
self.Us = self.Ubg[0]-... | python | def _initialize_stretching_matrix(self):
""" Set up the stretching matrix """
self.S = np.zeros((self.nz, self.nz))
if (self.nz==2) and (self.rd) and (self.delta):
self.del1 = self.delta/(self.delta+1.)
self.del2 = (self.delta+1.)**-1
self.Us = self.Ubg[0]-... | [
"def",
"_initialize_stretching_matrix",
"(",
"self",
")",
":",
"self",
".",
"S",
"=",
"np",
".",
"zeros",
"(",
"(",
"self",
".",
"nz",
",",
"self",
".",
"nz",
")",
")",
"if",
"(",
"self",
".",
"nz",
"==",
"2",
")",
"and",
"(",
"self",
".",
"rd"... | Set up the stretching matrix | [
"Set",
"up",
"the",
"stretching",
"matrix"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L130-L162 |
pyqg/pyqg | pyqg/layered_model.py | LayeredModel._initialize_background | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
self.H = self.Hi.sum()
if np.asarray(self.U).ndim == 2:
self.Ubg = self.U * np.ones((self.ny))
else:
self.Ubg = np.expand_dims(self.U,axis=1) * np.ones((self.ny))
... | python | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
self.H = self.Hi.sum()
if np.asarray(self.U).ndim == 2:
self.Ubg = self.U * np.ones((self.ny))
else:
self.Ubg = np.expand_dims(self.U,axis=1) * np.ones((self.ny))
... | [
"def",
"_initialize_background",
"(",
"self",
")",
":",
"self",
".",
"H",
"=",
"self",
".",
"Hi",
".",
"sum",
"(",
")",
"if",
"np",
".",
"asarray",
"(",
"self",
".",
"U",
")",
".",
"ndim",
"==",
"2",
":",
"self",
".",
"Ubg",
"=",
"self",
".",
... | Set up background state (zonal flow and PV gradients). | [
"Set",
"up",
"background",
"state",
"(",
"zonal",
"flow",
"and",
"PV",
"gradients",
")",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L164-L208 |
pyqg/pyqg | pyqg/layered_model.py | LayeredModel._calc_eddy_time | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = 0.
for j in range(self.nz):
ens = .5*self.Hi[j] * self.spec_var(self.wv2*self.ph[j])
return 2.*pi*np.sqrt( self.H / ens.sum() ) / 86400 | python | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = 0.
for j in range(self.nz):
ens = .5*self.Hi[j] * self.spec_var(self.wv2*self.ph[j])
return 2.*pi*np.sqrt( self.H / ens.sum() ) / 86400 | [
"def",
"_calc_eddy_time",
"(",
"self",
")",
":",
"ens",
"=",
"0.",
"for",
"j",
"in",
"range",
"(",
"self",
".",
"nz",
")",
":",
"ens",
"=",
".5",
"*",
"self",
".",
"Hi",
"[",
"j",
"]",
"*",
"self",
".",
"spec_var",
"(",
"self",
".",
"wv2",
"*... | estimate the eddy turn-over time in days | [
"estimate",
"the",
"eddy",
"turn",
"-",
"over",
"time",
"in",
"days"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L255-L261 |
pyqg/pyqg | pyqg/layered_model.py | LayeredModel._initialize_model_diagnostics | def _initialize_model_diagnostics(self):
""" Extra diagnostics for layered model """
self.add_diagnostic('entspec',
description='barotropic enstrophy spectrum',
function= (lambda self:
np.abs((self.Hi[:,np.newaxis,np.newaxis]*self.qh).sum(axis=0))**2/... | python | def _initialize_model_diagnostics(self):
""" Extra diagnostics for layered model """
self.add_diagnostic('entspec',
description='barotropic enstrophy spectrum',
function= (lambda self:
np.abs((self.Hi[:,np.newaxis,np.newaxis]*self.qh).sum(axis=0))**2/... | [
"def",
"_initialize_model_diagnostics",
"(",
"self",
")",
":",
"self",
".",
"add_diagnostic",
"(",
"'entspec'",
",",
"description",
"=",
"'barotropic enstrophy spectrum'",
",",
"function",
"=",
"(",
"lambda",
"self",
":",
"np",
".",
"abs",
"(",
"(",
"self",
".... | Extra diagnostics for layered model | [
"Extra",
"diagnostics",
"for",
"layered",
"model"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/layered_model.py#L277-L327 |
pyqg/pyqg | pyqg/qg_model.py | QGModel._initialize_background | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
# Background zonal flow (m/s):
self.H = self.Hi.sum()
self.set_U1U2(self.U1, self.U2)
self.U = self.U1 - self.U2
# the F parameters
self.F1 = self.rd**-2 / (1.+self.de... | python | def _initialize_background(self):
"""Set up background state (zonal flow and PV gradients)."""
# Background zonal flow (m/s):
self.H = self.Hi.sum()
self.set_U1U2(self.U1, self.U2)
self.U = self.U1 - self.U2
# the F parameters
self.F1 = self.rd**-2 / (1.+self.de... | [
"def",
"_initialize_background",
"(",
"self",
")",
":",
"# Background zonal flow (m/s):",
"self",
".",
"H",
"=",
"self",
".",
"Hi",
".",
"sum",
"(",
")",
"self",
".",
"set_U1U2",
"(",
"self",
".",
"U1",
",",
"self",
".",
"U2",
")",
"self",
".",
"U",
... | Set up background state (zonal flow and PV gradients). | [
"Set",
"up",
"background",
"state",
"(",
"zonal",
"flow",
"and",
"PV",
"gradients",
")",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L114-L142 |
pyqg/pyqg | pyqg/qg_model.py | QGModel.set_q1q2 | def set_q1q2(self, q1, q2, check=False):
"""Set upper and lower layer PV anomalies.
Parameters
----------
q1 : array-like
Upper layer PV anomaly in spatial coordinates.
q1 : array-like
Lower layer PV anomaly in spatial coordinates.
"""
se... | python | def set_q1q2(self, q1, q2, check=False):
"""Set upper and lower layer PV anomalies.
Parameters
----------
q1 : array-like
Upper layer PV anomaly in spatial coordinates.
q1 : array-like
Lower layer PV anomaly in spatial coordinates.
"""
se... | [
"def",
"set_q1q2",
"(",
"self",
",",
"q1",
",",
"q2",
",",
"check",
"=",
"False",
")",
":",
"self",
".",
"set_q",
"(",
"np",
".",
"vstack",
"(",
"[",
"q1",
"[",
"np",
".",
"newaxis",
",",
":",
",",
":",
"]",
",",
"q2",
"[",
"np",
".",
"newa... | Set upper and lower layer PV anomalies.
Parameters
----------
q1 : array-like
Upper layer PV anomaly in spatial coordinates.
q1 : array-like
Lower layer PV anomaly in spatial coordinates. | [
"Set",
"upper",
"and",
"lower",
"layer",
"PV",
"anomalies",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L170-L191 |
pyqg/pyqg | pyqg/qg_model.py | QGModel.set_U1U2 | def set_U1U2(self, U1, U2):
"""Set background zonal flow.
Parameters
----------
U1 : number
Upper layer flow. Units: m/s
U2 : number
Lower layer flow. Units: m/s
"""
if len(np.shape(U1)) == 0:
U1 = U1 * np.ones((self.ny))
... | python | def set_U1U2(self, U1, U2):
"""Set background zonal flow.
Parameters
----------
U1 : number
Upper layer flow. Units: m/s
U2 : number
Lower layer flow. Units: m/s
"""
if len(np.shape(U1)) == 0:
U1 = U1 * np.ones((self.ny))
... | [
"def",
"set_U1U2",
"(",
"self",
",",
"U1",
",",
"U2",
")",
":",
"if",
"len",
"(",
"np",
".",
"shape",
"(",
"U1",
")",
")",
"==",
"0",
":",
"U1",
"=",
"U1",
"*",
"np",
".",
"ones",
"(",
"(",
"self",
".",
"ny",
")",
")",
"if",
"len",
"(",
... | Set background zonal flow.
Parameters
----------
U1 : number
Upper layer flow. Units: m/s
U2 : number
Lower layer flow. Units: m/s | [
"Set",
"background",
"zonal",
"flow",
"."
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L193-L211 |
pyqg/pyqg | pyqg/qg_model.py | QGModel._calc_eddy_time | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = .5*self.Hi[0] * self.spec_var(self.wv2*self.ph1) + \
.5*self.Hi[1] * self.spec_var(self.wv2*self.ph2)
return 2.*pi*np.sqrt( self.H / ens ) / 86400 | python | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = .5*self.Hi[0] * self.spec_var(self.wv2*self.ph1) + \
.5*self.Hi[1] * self.spec_var(self.wv2*self.ph2)
return 2.*pi*np.sqrt( self.H / ens ) / 86400 | [
"def",
"_calc_eddy_time",
"(",
"self",
")",
":",
"ens",
"=",
".5",
"*",
"self",
".",
"Hi",
"[",
"0",
"]",
"*",
"self",
".",
"spec_var",
"(",
"self",
".",
"wv2",
"*",
"self",
".",
"ph1",
")",
"+",
".5",
"*",
"self",
".",
"Hi",
"[",
"1",
"]",
... | estimate the eddy turn-over time in days | [
"estimate",
"the",
"eddy",
"turn",
"-",
"over",
"time",
"in",
"days"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L228-L234 |
pyqg/pyqg | pyqg/qg_model.py | QGModel._initialize_model_diagnostics | def _initialize_model_diagnostics(self):
"""Extra diagnostics for two-layer model"""
self.add_diagnostic('entspec',
description='barotropic enstrophy spectrum',
function= (lambda self:
np.abs(self.del1*self.qh[0] + self.del2*self.qh[1])**2.)
)
... | python | def _initialize_model_diagnostics(self):
"""Extra diagnostics for two-layer model"""
self.add_diagnostic('entspec',
description='barotropic enstrophy spectrum',
function= (lambda self:
np.abs(self.del1*self.qh[0] + self.del2*self.qh[1])**2.)
)
... | [
"def",
"_initialize_model_diagnostics",
"(",
"self",
")",
":",
"self",
".",
"add_diagnostic",
"(",
"'entspec'",
",",
"description",
"=",
"'barotropic enstrophy spectrum'",
",",
"function",
"=",
"(",
"lambda",
"self",
":",
"np",
".",
"abs",
"(",
"self",
".",
"d... | Extra diagnostics for two-layer model | [
"Extra",
"diagnostics",
"for",
"two",
"-",
"layer",
"model"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/qg_model.py#L246-L286 |
pyqg/pyqg | pyqg/bt_model.py | BTModel._initialize_inversion_matrix | def _initialize_inversion_matrix(self):
""" the inversion """
# The bt model is diagonal. The inversion is simply qh = -kappa**2 ph
self.a = -(self.wv2i+self.kd2)[np.newaxis, np.newaxis, :, :] | python | def _initialize_inversion_matrix(self):
""" the inversion """
# The bt model is diagonal. The inversion is simply qh = -kappa**2 ph
self.a = -(self.wv2i+self.kd2)[np.newaxis, np.newaxis, :, :] | [
"def",
"_initialize_inversion_matrix",
"(",
"self",
")",
":",
"# The bt model is diagonal. The inversion is simply qh = -kappa**2 ph",
"self",
".",
"a",
"=",
"-",
"(",
"self",
".",
"wv2i",
"+",
"self",
".",
"kd2",
")",
"[",
"np",
".",
"newaxis",
",",
"np",
".",
... | the inversion | [
"the",
"inversion"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/bt_model.py#L75-L78 |
pyqg/pyqg | pyqg/bt_model.py | BTModel._calc_eddy_time | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = .5*self.H * self.spec_var(self.wv2*self.ph)
return 2.*pi*np.sqrt( self.H / ens ) / year | python | def _calc_eddy_time(self):
""" estimate the eddy turn-over time in days """
ens = .5*self.H * self.spec_var(self.wv2*self.ph)
return 2.*pi*np.sqrt( self.H / ens ) / year | [
"def",
"_calc_eddy_time",
"(",
"self",
")",
":",
"ens",
"=",
".5",
"*",
"self",
".",
"H",
"*",
"self",
".",
"spec_var",
"(",
"self",
".",
"wv2",
"*",
"self",
".",
"ph",
")",
"return",
"2.",
"*",
"pi",
"*",
"np",
".",
"sqrt",
"(",
"self",
".",
... | estimate the eddy turn-over time in days | [
"estimate",
"the",
"eddy",
"turn",
"-",
"over",
"time",
"in",
"days"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/bt_model.py#L123-L126 |
pyqg/pyqg | pyqg/point_vortex.py | PointVortexArray2D.calc_uv | def calc_uv(self, x, y, prev=False):
"""Calculate velocity at x and y points due to vortex velocity field.
Assumes x and y are vortex positions and are ordered the same as
x0 and y0. The ordering is used to neglect to vortex self interaction."""
assert len(x) == self.N
assert len... | python | def calc_uv(self, x, y, prev=False):
"""Calculate velocity at x and y points due to vortex velocity field.
Assumes x and y are vortex positions and are ordered the same as
x0 and y0. The ordering is used to neglect to vortex self interaction."""
assert len(x) == self.N
assert len... | [
"def",
"calc_uv",
"(",
"self",
",",
"x",
",",
"y",
",",
"prev",
"=",
"False",
")",
":",
"assert",
"len",
"(",
"x",
")",
"==",
"self",
".",
"N",
"assert",
"len",
"(",
"y",
")",
"==",
"self",
".",
"N",
"u",
"=",
"np",
".",
"zeros",
"(",
"self... | Calculate velocity at x and y points due to vortex velocity field.
Assumes x and y are vortex positions and are ordered the same as
x0 and y0. The ordering is used to neglect to vortex self interaction. | [
"Calculate",
"velocity",
"at",
"x",
"and",
"y",
"points",
"due",
"to",
"vortex",
"velocity",
"field",
".",
"Assumes",
"x",
"and",
"y",
"are",
"vortex",
"positions",
"and",
"are",
"ordered",
"the",
"same",
"as",
"x0",
"and",
"y0",
".",
"The",
"ordering",
... | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/point_vortex.py#L32-L52 |
pyqg/pyqg | pyqg/point_vortex.py | PointVortexArray2D.uv_at_xy | def uv_at_xy(self, x, y, x0, y0, s0):
"""Returns two arrays of u, v"""
dx, dy = self.distance(x0, y0, x, y)
#print 'dx, dy:', dx, dy
rr2 = (dx**2 + dy**2)**-1
u = - s0 * dy * r_twopi * rr2
v = s0 * dx * r_twopi * rr2
#print 'u, v', u, v
return u, v | python | def uv_at_xy(self, x, y, x0, y0, s0):
"""Returns two arrays of u, v"""
dx, dy = self.distance(x0, y0, x, y)
#print 'dx, dy:', dx, dy
rr2 = (dx**2 + dy**2)**-1
u = - s0 * dy * r_twopi * rr2
v = s0 * dx * r_twopi * rr2
#print 'u, v', u, v
return u, v | [
"def",
"uv_at_xy",
"(",
"self",
",",
"x",
",",
"y",
",",
"x0",
",",
"y0",
",",
"s0",
")",
":",
"dx",
",",
"dy",
"=",
"self",
".",
"distance",
"(",
"x0",
",",
"y0",
",",
"x",
",",
"y",
")",
"#print 'dx, dy:', dx, dy",
"rr2",
"=",
"(",
"dx",
"*... | Returns two arrays of u, v | [
"Returns",
"two",
"arrays",
"of",
"u",
"v"
] | train | https://github.com/pyqg/pyqg/blob/4f41584a12bcbf8657785b8cb310fa5065ecabd1/pyqg/point_vortex.py#L54-L62 |
brentp/interlap | interlap.py | reduce | def reduce(args):
"""
>>> reduce([(2, 4), (4, 9)])
[(2, 4), (4, 9)]
>>> reduce([(2, 6), (4, 10)])
[(2, 10)]
"""
if len(args) < 2: return args
args.sort()
ret = [args[0]]
for next_i, (s, e) in enumerate(args, start=1):
if next_i == len(args):
ret[-1] = ret[-1]... | python | def reduce(args):
"""
>>> reduce([(2, 4), (4, 9)])
[(2, 4), (4, 9)]
>>> reduce([(2, 6), (4, 10)])
[(2, 10)]
"""
if len(args) < 2: return args
args.sort()
ret = [args[0]]
for next_i, (s, e) in enumerate(args, start=1):
if next_i == len(args):
ret[-1] = ret[-1]... | [
"def",
"reduce",
"(",
"args",
")",
":",
"if",
"len",
"(",
"args",
")",
"<",
"2",
":",
"return",
"args",
"args",
".",
"sort",
"(",
")",
"ret",
"=",
"[",
"args",
"[",
"0",
"]",
"]",
"for",
"next_i",
",",
"(",
"s",
",",
"e",
")",
"in",
"enumer... | >>> reduce([(2, 4), (4, 9)])
[(2, 4), (4, 9)]
>>> reduce([(2, 6), (4, 10)])
[(2, 10)] | [
">>>",
"reduce",
"(",
"[",
"(",
"2",
"4",
")",
"(",
"4",
"9",
")",
"]",
")",
"[",
"(",
"2",
"4",
")",
"(",
"4",
"9",
")",
"]"
] | train | https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L224-L245 |
brentp/interlap | interlap.py | InterLap.add | def add(self, ranges):
r"""Add a single (or many) [start, end, \*] item to the tree."""
if len(ranges) and isinstance(ranges[0], int_types):
ranges = [ranges]
iset = self._iset
self._maxlen = max(self._maxlen, max(r[1] - r[0] + 1 for r in ranges))
if len(ranges) > 30... | python | def add(self, ranges):
r"""Add a single (or many) [start, end, \*] item to the tree."""
if len(ranges) and isinstance(ranges[0], int_types):
ranges = [ranges]
iset = self._iset
self._maxlen = max(self._maxlen, max(r[1] - r[0] + 1 for r in ranges))
if len(ranges) > 30... | [
"def",
"add",
"(",
"self",
",",
"ranges",
")",
":",
"if",
"len",
"(",
"ranges",
")",
"and",
"isinstance",
"(",
"ranges",
"[",
"0",
"]",
",",
"int_types",
")",
":",
"ranges",
"=",
"[",
"ranges",
"]",
"iset",
"=",
"self",
".",
"_iset",
"self",
".",... | r"""Add a single (or many) [start, end, \*] item to the tree. | [
"r",
"Add",
"a",
"single",
"(",
"or",
"many",
")",
"[",
"start",
"end",
"\\",
"*",
"]",
"item",
"to",
"the",
"tree",
"."
] | train | https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L133-L145 |
brentp/interlap | interlap.py | InterLap.find | def find(self, other):
"""Return an interable of elements that overlap other in the tree."""
iset = self._iset
l = binsearch_left_start(iset, other[0] - self._maxlen, 0, len(iset))
r = binsearch_right_end(iset, other[1], 0, len(iset))
iopts = iset[l:r]
iiter = (s for s in... | python | def find(self, other):
"""Return an interable of elements that overlap other in the tree."""
iset = self._iset
l = binsearch_left_start(iset, other[0] - self._maxlen, 0, len(iset))
r = binsearch_right_end(iset, other[1], 0, len(iset))
iopts = iset[l:r]
iiter = (s for s in... | [
"def",
"find",
"(",
"self",
",",
"other",
")",
":",
"iset",
"=",
"self",
".",
"_iset",
"l",
"=",
"binsearch_left_start",
"(",
"iset",
",",
"other",
"[",
"0",
"]",
"-",
"self",
".",
"_maxlen",
",",
"0",
",",
"len",
"(",
"iset",
")",
")",
"r",
"=... | Return an interable of elements that overlap other in the tree. | [
"Return",
"an",
"interable",
"of",
"elements",
"that",
"overlap",
"other",
"in",
"the",
"tree",
"."
] | train | https://github.com/brentp/interlap/blob/3c4a5923c97a5d9a11571e0c9ea5bb7ea4e784ee/interlap.py#L153-L160 |
gumblex/zhconv | zhconv/zhconv.py | loaddict | def loaddict(filename=DICTIONARY):
"""
Load the dictionary from a specific JSON file.
"""
global zhcdicts
if zhcdicts:
return
if filename == _DEFAULT_DICT:
zhcdicts = json.loads(get_module_res(filename).read().decode('utf-8'))
else:
with open(filename, 'rb') as f:
... | python | def loaddict(filename=DICTIONARY):
"""
Load the dictionary from a specific JSON file.
"""
global zhcdicts
if zhcdicts:
return
if filename == _DEFAULT_DICT:
zhcdicts = json.loads(get_module_res(filename).read().decode('utf-8'))
else:
with open(filename, 'rb') as f:
... | [
"def",
"loaddict",
"(",
"filename",
"=",
"DICTIONARY",
")",
":",
"global",
"zhcdicts",
"if",
"zhcdicts",
":",
"return",
"if",
"filename",
"==",
"_DEFAULT_DICT",
":",
"zhcdicts",
"=",
"json",
".",
"loads",
"(",
"get_module_res",
"(",
"filename",
")",
".",
"... | Load the dictionary from a specific JSON file. | [
"Load",
"the",
"dictionary",
"from",
"a",
"specific",
"JSON",
"file",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L68-L81 |
gumblex/zhconv | zhconv/zhconv.py | getdict | def getdict(locale):
"""
Generate or get convertion dict cache for certain locale.
Dictionaries are loaded on demand.
"""
global zhcdicts, dict_zhcn, dict_zhsg, dict_zhtw, dict_zhhk, pfsdict
if zhcdicts is None:
loaddict(DICTIONARY)
if locale == 'zh-cn':
if dict_zhcn:
... | python | def getdict(locale):
"""
Generate or get convertion dict cache for certain locale.
Dictionaries are loaded on demand.
"""
global zhcdicts, dict_zhcn, dict_zhsg, dict_zhtw, dict_zhhk, pfsdict
if zhcdicts is None:
loaddict(DICTIONARY)
if locale == 'zh-cn':
if dict_zhcn:
... | [
"def",
"getdict",
"(",
"locale",
")",
":",
"global",
"zhcdicts",
",",
"dict_zhcn",
",",
"dict_zhsg",
",",
"dict_zhtw",
",",
"dict_zhhk",
",",
"pfsdict",
"if",
"zhcdicts",
"is",
"None",
":",
"loaddict",
"(",
"DICTIONARY",
")",
"if",
"locale",
"==",
"'zh-cn'... | Generate or get convertion dict cache for certain locale.
Dictionaries are loaded on demand. | [
"Generate",
"or",
"get",
"convertion",
"dict",
"cache",
"for",
"certain",
"locale",
".",
"Dictionaries",
"are",
"loaded",
"on",
"demand",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L83-L127 |
gumblex/zhconv | zhconv/zhconv.py | issimp | def issimp(s, full=False):
"""
Detect text is whether Simplified Chinese or Traditional Chinese.
Returns True for Simplified; False for Traditional; None for unknown.
If full=False, it returns once first simplified- or traditional-only
character is encountered, so it's for quick and rough identifica... | python | def issimp(s, full=False):
"""
Detect text is whether Simplified Chinese or Traditional Chinese.
Returns True for Simplified; False for Traditional; None for unknown.
If full=False, it returns once first simplified- or traditional-only
character is encountered, so it's for quick and rough identifica... | [
"def",
"issimp",
"(",
"s",
",",
"full",
"=",
"False",
")",
":",
"if",
"zhcdicts",
"is",
"None",
":",
"loaddict",
"(",
"DICTIONARY",
")",
"simp",
",",
"trad",
"=",
"0",
",",
"0",
"if",
"full",
":",
"for",
"ch",
"in",
"s",
":",
"if",
"ch",
"in",
... | Detect text is whether Simplified Chinese or Traditional Chinese.
Returns True for Simplified; False for Traditional; None for unknown.
If full=False, it returns once first simplified- or traditional-only
character is encountered, so it's for quick and rough identification;
else, it compares the count a... | [
"Detect",
"text",
"is",
"whether",
"Simplified",
"Chinese",
"or",
"Traditional",
"Chinese",
".",
"Returns",
"True",
"for",
"Simplified",
";",
"False",
"for",
"Traditional",
";",
"None",
"for",
"unknown",
".",
"If",
"full",
"=",
"False",
"it",
"returns",
"onc... | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L136-L168 |
gumblex/zhconv | zhconv/zhconv.py | convtable2dict | def convtable2dict(convtable, locale, update=None):
"""
Convert a list of conversion dict to a dict for a certain locale.
>>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items())
[('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('... | python | def convtable2dict(convtable, locale, update=None):
"""
Convert a list of conversion dict to a dict for a certain locale.
>>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items())
[('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('... | [
"def",
"convtable2dict",
"(",
"convtable",
",",
"locale",
",",
"update",
"=",
"None",
")",
":",
"rdict",
"=",
"update",
".",
"copy",
"(",
")",
"if",
"update",
"else",
"{",
"}",
"for",
"r",
"in",
"convtable",
":",
"if",
"':uni'",
"in",
"r",
":",
"if... | Convert a list of conversion dict to a dict for a certain locale.
>>> sorted(convtable2dict([{'zh-hk': '列斯', 'zh-hans': '利兹', 'zh': '利兹', 'zh-tw': '里茲'}, {':uni': '巨集', 'zh-cn': '宏'}], 'zh-cn').items())
[('列斯', '利兹'), ('利兹', '利兹'), ('巨集', '宏'), ('里茲', '利兹')] | [
"Convert",
"a",
"list",
"of",
"conversion",
"dict",
"to",
"a",
"dict",
"for",
"a",
"certain",
"locale",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L176-L196 |
gumblex/zhconv | zhconv/zhconv.py | tokenize | def tokenize(s, locale, update=None):
"""
Tokenize `s` according to corresponding locale dictionary.
Don't use this for serious text processing.
"""
zhdict = getdict(locale)
pfset = pfsdict[locale]
if update:
zhdict = zhdict.copy()
zhdict.update(update)
newset = set()... | python | def tokenize(s, locale, update=None):
"""
Tokenize `s` according to corresponding locale dictionary.
Don't use this for serious text processing.
"""
zhdict = getdict(locale)
pfset = pfsdict[locale]
if update:
zhdict = zhdict.copy()
zhdict.update(update)
newset = set()... | [
"def",
"tokenize",
"(",
"s",
",",
"locale",
",",
"update",
"=",
"None",
")",
":",
"zhdict",
"=",
"getdict",
"(",
"locale",
")",
"pfset",
"=",
"pfsdict",
"[",
"locale",
"]",
"if",
"update",
":",
"zhdict",
"=",
"zhdict",
".",
"copy",
"(",
")",
"zhdic... | Tokenize `s` according to corresponding locale dictionary.
Don't use this for serious text processing. | [
"Tokenize",
"s",
"according",
"to",
"corresponding",
"locale",
"dictionary",
".",
"Don",
"t",
"use",
"this",
"for",
"serious",
"text",
"processing",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L198-L233 |
gumblex/zhconv | zhconv/zhconv.py | convert_for_mw | def convert_for_mw(s, locale, update=None):
"""
Recognizes MediaWiki's human conversion format.
Use locale='zh' for no conversion.
Reference: (all tests passed)
https://zh.wikipedia.org/wiki/Help:高级字词转换语法
https://www.mediawiki.org/wiki/Writing_systems/Syntax
>>> print(convert_for_mw('在现代,机... | python | def convert_for_mw(s, locale, update=None):
"""
Recognizes MediaWiki's human conversion format.
Use locale='zh' for no conversion.
Reference: (all tests passed)
https://zh.wikipedia.org/wiki/Help:高级字词转换语法
https://www.mediawiki.org/wiki/Writing_systems/Syntax
>>> print(convert_for_mw('在现代,机... | [
"def",
"convert_for_mw",
"(",
"s",
",",
"locale",
",",
"update",
"=",
"None",
")",
":",
"ch",
"=",
"[",
"]",
"rules",
"=",
"[",
"]",
"ruledict",
"=",
"update",
".",
"copy",
"(",
")",
"if",
"update",
"else",
"{",
"}",
"nested",
"=",
"0",
"block",
... | Recognizes MediaWiki's human conversion format.
Use locale='zh' for no conversion.
Reference: (all tests passed)
https://zh.wikipedia.org/wiki/Help:高级字词转换语法
https://www.mediawiki.org/wiki/Writing_systems/Syntax
>>> print(convert_for_mw('在现代,机械计算-{}-机的应用已经完全被电子计算-{}-机所取代', 'zh-hk'))
在現代,機械計算機的應... | [
"Recognizes",
"MediaWiki",
"s",
"human",
"conversion",
"format",
".",
"Use",
"locale",
"=",
"zh",
"for",
"no",
"conversion",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L292-L425 |
gumblex/zhconv | zhconv/zhconv.py | main | def main():
"""
Simple stdin/stdout interface.
"""
if len(sys.argv) == 2 and sys.argv[1] in Locales:
locale = sys.argv[1]
convertfunc = convert
elif len(sys.argv) == 3 and sys.argv[1] == '-w' and sys.argv[2] in Locales:
locale = sys.argv[2]
convertfunc = convert_for_m... | python | def main():
"""
Simple stdin/stdout interface.
"""
if len(sys.argv) == 2 and sys.argv[1] in Locales:
locale = sys.argv[1]
convertfunc = convert
elif len(sys.argv) == 3 and sys.argv[1] == '-w' and sys.argv[2] in Locales:
locale = sys.argv[2]
convertfunc = convert_for_m... | [
"def",
"main",
"(",
")",
":",
"if",
"len",
"(",
"sys",
".",
"argv",
")",
"==",
"2",
"and",
"sys",
".",
"argv",
"[",
"1",
"]",
"in",
"Locales",
":",
"locale",
"=",
"sys",
".",
"argv",
"[",
"1",
"]",
"convertfunc",
"=",
"convert",
"elif",
"len",
... | Simple stdin/stdout interface. | [
"Simple",
"stdin",
"/",
"stdout",
"interface",
"."
] | train | https://github.com/gumblex/zhconv/blob/925c0f9494f3439bc05526e7e89bb5f0ab3d185e/zhconv/zhconv.py#L449-L475 |
glasslion/django-qiniu-storage | qiniustorage/utils.py | bucket_lister | def bucket_lister(manager, bucket_name, prefix=None, marker=None, limit=None):
"""
A generator function for listing keys in a bucket.
"""
eof = False
while not eof:
ret, eof, info = manager.list(bucket_name, prefix=prefix, limit=limit,
marker=marker)
... | python | def bucket_lister(manager, bucket_name, prefix=None, marker=None, limit=None):
"""
A generator function for listing keys in a bucket.
"""
eof = False
while not eof:
ret, eof, info = manager.list(bucket_name, prefix=prefix, limit=limit,
marker=marker)
... | [
"def",
"bucket_lister",
"(",
"manager",
",",
"bucket_name",
",",
"prefix",
"=",
"None",
",",
"marker",
"=",
"None",
",",
"limit",
"=",
"None",
")",
":",
"eof",
"=",
"False",
"while",
"not",
"eof",
":",
"ret",
",",
"eof",
",",
"info",
"=",
"manager",
... | A generator function for listing keys in a bucket. | [
"A",
"generator",
"function",
"for",
"listing",
"keys",
"in",
"a",
"bucket",
"."
] | train | https://github.com/glasslion/django-qiniu-storage/blob/b046ec0b67ebcf8cd9eb09c60f7db4a7e4fab7ad/qiniustorage/utils.py#L17-L31 |
glasslion/django-qiniu-storage | qiniustorage/backends.py | get_qiniu_config | def get_qiniu_config(name, default=None):
"""
Get configuration variable from environment variable
or django setting.py
"""
config = os.environ.get(name, getattr(settings, name, default))
if config is not None:
if isinstance(config, six.string_types):
return config.strip()
... | python | def get_qiniu_config(name, default=None):
"""
Get configuration variable from environment variable
or django setting.py
"""
config = os.environ.get(name, getattr(settings, name, default))
if config is not None:
if isinstance(config, six.string_types):
return config.strip()
... | [
"def",
"get_qiniu_config",
"(",
"name",
",",
"default",
"=",
"None",
")",
":",
"config",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"name",
",",
"getattr",
"(",
"settings",
",",
"name",
",",
"default",
")",
")",
"if",
"config",
"is",
"not",
"None",
... | Get configuration variable from environment variable
or django setting.py | [
"Get",
"configuration",
"variable",
"from",
"environment",
"variable",
"or",
"django",
"setting",
".",
"py"
] | train | https://github.com/glasslion/django-qiniu-storage/blob/b046ec0b67ebcf8cd9eb09c60f7db4a7e4fab7ad/qiniustorage/backends.py#L27-L41 |
non-Jedi/gyr | gyr/api.py | MatrixASHttpAPI.register | def register(self, username=""):
"""Performs /register with type: m.login.application_service
Args:
username(str): Username to register.
"""
if not username:
username = utils.mxid2localpart(self.identity)
content = {
"type": "m.login.applicati... | python | def register(self, username=""):
"""Performs /register with type: m.login.application_service
Args:
username(str): Username to register.
"""
if not username:
username = utils.mxid2localpart(self.identity)
content = {
"type": "m.login.applicati... | [
"def",
"register",
"(",
"self",
",",
"username",
"=",
"\"\"",
")",
":",
"if",
"not",
"username",
":",
"username",
"=",
"utils",
".",
"mxid2localpart",
"(",
"self",
".",
"identity",
")",
"content",
"=",
"{",
"\"type\"",
":",
"\"m.login.application_service\"",... | Performs /register with type: m.login.application_service
Args:
username(str): Username to register. | [
"Performs",
"/",
"register",
"with",
"type",
":",
"m",
".",
"login",
".",
"application_service"
] | train | https://github.com/non-Jedi/gyr/blob/9f7bfe033b9d3bbfd3a9e8aea02e35526b53125e/gyr/api.py#L58-L71 |
ocaballeror/LyricFetch | lyricfetch/cli.py | load_from_file | def load_from_file(filename):
"""
Load a list of filenames from an external text file.
"""
if os.path.isdir(filename):
logger.error("Err: File '%s' is a directory", filename)
return None
if not os.path.isfile(filename):
logger.error("Err: File '%s' does not exist", filename)
... | python | def load_from_file(filename):
"""
Load a list of filenames from an external text file.
"""
if os.path.isdir(filename):
logger.error("Err: File '%s' is a directory", filename)
return None
if not os.path.isfile(filename):
logger.error("Err: File '%s' does not exist", filename)
... | [
"def",
"load_from_file",
"(",
"filename",
")",
":",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"filename",
")",
":",
"logger",
".",
"error",
"(",
"\"Err: File '%s' is a directory\"",
",",
"filename",
")",
"return",
"None",
"if",
"not",
"os",
".",
"path",
... | Load a list of filenames from an external text file. | [
"Load",
"a",
"list",
"of",
"filenames",
"from",
"an",
"external",
"text",
"file",
"."
] | train | https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L17-L35 |
ocaballeror/LyricFetch | lyricfetch/cli.py | parse_argv | def parse_argv():
"""
Parse command line arguments. Settings will be stored in the global
variables declared above.
"""
parser = argparse.ArgumentParser(description='Find lyrics for a set of mp3'
' files and embed them as metadata')
parser.add_argument('-j', ... | python | def parse_argv():
"""
Parse command line arguments. Settings will be stored in the global
variables declared above.
"""
parser = argparse.ArgumentParser(description='Find lyrics for a set of mp3'
' files and embed them as metadata')
parser.add_argument('-j', ... | [
"def",
"parse_argv",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"'Find lyrics for a set of mp3'",
"' files and embed them as metadata'",
")",
"parser",
".",
"add_argument",
"(",
"'-j'",
",",
"'--jobs'",
",",
"help",
"="... | Parse command line arguments. Settings will be stored in the global
variables declared above. | [
"Parse",
"command",
"line",
"arguments",
".",
"Settings",
"will",
"be",
"stored",
"in",
"the",
"global",
"variables",
"declared",
"above",
"."
] | train | https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L38-L99 |
ocaballeror/LyricFetch | lyricfetch/cli.py | main | def main():
"""
Main function.
"""
msg = ''
try:
songs = parse_argv()
if not songs:
msg = 'No songs specified'
except ValueError as error:
msg = str(error)
if msg:
logger.error('%s: Error: %s', sys.argv[0], msg)
return 1
logger.debug('... | python | def main():
"""
Main function.
"""
msg = ''
try:
songs = parse_argv()
if not songs:
msg = 'No songs specified'
except ValueError as error:
msg = str(error)
if msg:
logger.error('%s: Error: %s', sys.argv[0], msg)
return 1
logger.debug('... | [
"def",
"main",
"(",
")",
":",
"msg",
"=",
"''",
"try",
":",
"songs",
"=",
"parse_argv",
"(",
")",
"if",
"not",
"songs",
":",
"msg",
"=",
"'No songs specified'",
"except",
"ValueError",
"as",
"error",
":",
"msg",
"=",
"str",
"(",
"error",
")",
"if",
... | Main function. | [
"Main",
"function",
"."
] | train | https://github.com/ocaballeror/LyricFetch/blob/86e62fb39c4c413ad7e1acf5bf0d28c9ed7c8fcb/lyricfetch/cli.py#L102-L124 |
taskcluster/slugid.py | slugid/slugid.py | decode | def decode(slug):
"""
Returns the uuid.UUID object represented by the given v4 or "nice" slug
"""
if sys.version_info.major != 2 and isinstance(slug, bytes):
slug = slug.decode('ascii')
slug = slug + '==' # base64 padding
return uuid.UUID(bytes=base64.urlsafe_b64decode(slug)) | python | def decode(slug):
"""
Returns the uuid.UUID object represented by the given v4 or "nice" slug
"""
if sys.version_info.major != 2 and isinstance(slug, bytes):
slug = slug.decode('ascii')
slug = slug + '==' # base64 padding
return uuid.UUID(bytes=base64.urlsafe_b64decode(slug)) | [
"def",
"decode",
"(",
"slug",
")",
":",
"if",
"sys",
".",
"version_info",
".",
"major",
"!=",
"2",
"and",
"isinstance",
"(",
"slug",
",",
"bytes",
")",
":",
"slug",
"=",
"slug",
".",
"decode",
"(",
"'ascii'",
")",
"slug",
"=",
"slug",
"+",
"'=='",
... | Returns the uuid.UUID object represented by the given v4 or "nice" slug | [
"Returns",
"the",
"uuid",
".",
"UUID",
"object",
"represented",
"by",
"the",
"given",
"v4",
"or",
"nice",
"slug"
] | train | https://github.com/taskcluster/slugid.py/blob/7c2c58e79d8684a54c578302ad60b384e52bb09b/slugid/slugid.py#L24-L31 |
taskcluster/slugid.py | slugid/slugid.py | nice | def nice():
"""
Returns a randomly generated uuid v4 compliant slug which conforms to a set
of "nice" properties, at the cost of some entropy. Currently this means one
extra fixed bit (the first bit of the uuid is set to 0) which guarantees the
slug will begin with [A-Za-f]. For example such slugs d... | python | def nice():
"""
Returns a randomly generated uuid v4 compliant slug which conforms to a set
of "nice" properties, at the cost of some entropy. Currently this means one
extra fixed bit (the first bit of the uuid is set to 0) which guarantees the
slug will begin with [A-Za-f]. For example such slugs d... | [
"def",
"nice",
"(",
")",
":",
"rawBytes",
"=",
"bytearray",
"(",
"uuid",
".",
"uuid4",
"(",
")",
".",
"bytes",
")",
"rawBytes",
"[",
"0",
"]",
"=",
"rawBytes",
"[",
"0",
"]",
"&",
"0x7f",
"# Ensure slug starts with [A-Za-f]",
"return",
"_convert_bytes_to_s... | Returns a randomly generated uuid v4 compliant slug which conforms to a set
of "nice" properties, at the cost of some entropy. Currently this means one
extra fixed bit (the first bit of the uuid is set to 0) which guarantees the
slug will begin with [A-Za-f]. For example such slugs don't require special
... | [
"Returns",
"a",
"randomly",
"generated",
"uuid",
"v4",
"compliant",
"slug",
"which",
"conforms",
"to",
"a",
"set",
"of",
"nice",
"properties",
"at",
"the",
"cost",
"of",
"some",
"entropy",
".",
"Currently",
"this",
"means",
"one",
"extra",
"fixed",
"bit",
... | train | https://github.com/taskcluster/slugid.py/blob/7c2c58e79d8684a54c578302ad60b384e52bb09b/slugid/slugid.py#L41-L55 |
inodb/sufam | sufam/mutation.py | MutationsAtSinglePosition.filter_against_normal | def filter_against_normal(self, normal_mutations, maf_min=0.2,
maf_count_threshold=20, count_min=1):
"""Filters mutations that are in the given normal"""
assert(normal_mutations.chrom == self.chrom)
assert(normal_mutations.pos == self.pos)
assert(normal_muta... | python | def filter_against_normal(self, normal_mutations, maf_min=0.2,
maf_count_threshold=20, count_min=1):
"""Filters mutations that are in the given normal"""
assert(normal_mutations.chrom == self.chrom)
assert(normal_mutations.pos == self.pos)
assert(normal_muta... | [
"def",
"filter_against_normal",
"(",
"self",
",",
"normal_mutations",
",",
"maf_min",
"=",
"0.2",
",",
"maf_count_threshold",
"=",
"20",
",",
"count_min",
"=",
"1",
")",
":",
"assert",
"(",
"normal_mutations",
".",
"chrom",
"==",
"self",
".",
"chrom",
")",
... | Filters mutations that are in the given normal | [
"Filters",
"mutations",
"that",
"are",
"in",
"the",
"given",
"normal"
] | train | https://github.com/inodb/sufam/blob/d4e41c5478ca9ba58be44d95106885c096c90a74/sufam/mutation.py#L55-L81 |
inodb/sufam | sufam/mutation.py | Mutation.to_oncotator | def to_oncotator(self):
"""Returns mutation in oncotator input format. Assumes mutations have
vcf/mpileup style positions."""
if self.type == ".":
ref = self.ref
alt = self.change
start = self.pos
end = self.pos
elif self.type == "-":
... | python | def to_oncotator(self):
"""Returns mutation in oncotator input format. Assumes mutations have
vcf/mpileup style positions."""
if self.type == ".":
ref = self.ref
alt = self.change
start = self.pos
end = self.pos
elif self.type == "-":
... | [
"def",
"to_oncotator",
"(",
"self",
")",
":",
"if",
"self",
".",
"type",
"==",
"\".\"",
":",
"ref",
"=",
"self",
".",
"ref",
"alt",
"=",
"self",
".",
"change",
"start",
"=",
"self",
".",
"pos",
"end",
"=",
"self",
".",
"pos",
"elif",
"self",
".",... | Returns mutation in oncotator input format. Assumes mutations have
vcf/mpileup style positions. | [
"Returns",
"mutation",
"in",
"oncotator",
"input",
"format",
".",
"Assumes",
"mutations",
"have",
"vcf",
"/",
"mpileup",
"style",
"positions",
"."
] | train | https://github.com/inodb/sufam/blob/d4e41c5478ca9ba58be44d95106885c096c90a74/sufam/mutation.py#L116-L137 |
non-Jedi/gyr | gyr/server.py | Application.add_handlers | def add_handlers(self, room_handler=None, transaction_handler=None,
user_handler=None):
"""Adds routes to Application that use specified handlers."""
# Add all the normal matrix API routes
if room_handler:
room = resources.Room(room_handler,
... | python | def add_handlers(self, room_handler=None, transaction_handler=None,
user_handler=None):
"""Adds routes to Application that use specified handlers."""
# Add all the normal matrix API routes
if room_handler:
room = resources.Room(room_handler,
... | [
"def",
"add_handlers",
"(",
"self",
",",
"room_handler",
"=",
"None",
",",
"transaction_handler",
"=",
"None",
",",
"user_handler",
"=",
"None",
")",
":",
"# Add all the normal matrix API routes",
"if",
"room_handler",
":",
"room",
"=",
"resources",
".",
"Room",
... | Adds routes to Application that use specified handlers. | [
"Adds",
"routes",
"to",
"Application",
"that",
"use",
"specified",
"handlers",
"."
] | train | https://github.com/non-Jedi/gyr/blob/9f7bfe033b9d3bbfd3a9e8aea02e35526b53125e/gyr/server.py#L34-L49 |
HolmesNL/confidence | confidence/utils.py | _merge | def _merge(left, right, path=None, conflict=_Conflict.error):
"""
Merges values in place from *right* into *left*.
:param left: mapping to merge into
:param right: mapping to merge from
:param path: `list` of keys processed before (used for error reporting
only, should only need to be provi... | python | def _merge(left, right, path=None, conflict=_Conflict.error):
"""
Merges values in place from *right* into *left*.
:param left: mapping to merge into
:param right: mapping to merge from
:param path: `list` of keys processed before (used for error reporting
only, should only need to be provi... | [
"def",
"_merge",
"(",
"left",
",",
"right",
",",
"path",
"=",
"None",
",",
"conflict",
"=",
"_Conflict",
".",
"error",
")",
":",
"path",
"=",
"path",
"or",
"[",
"]",
"conflict",
"=",
"_Conflict",
"(",
"conflict",
")",
"for",
"key",
"in",
"right",
"... | Merges values in place from *right* into *left*.
:param left: mapping to merge into
:param right: mapping to merge from
:param path: `list` of keys processed before (used for error reporting
only, should only need to be provided by recursive calls)
:param conflict: action to be taken on merge c... | [
"Merges",
"values",
"in",
"place",
"from",
"*",
"right",
"*",
"into",
"*",
"left",
"*",
"."
] | train | https://github.com/HolmesNL/confidence/blob/e14d2d8769a01fa55676716f7a2f22714c2616d3/confidence/utils.py#L13-L46 |
HolmesNL/confidence | confidence/utils.py | _split_keys | def _split_keys(mapping, separator='.', colliding=None):
"""
Recursively walks *mapping* to split keys that contain the separator into
nested mappings.
.. note::
Keys not of type `str` are not supported and will raise errors.
:param mapping: the mapping to process
:param separator: th... | python | def _split_keys(mapping, separator='.', colliding=None):
"""
Recursively walks *mapping* to split keys that contain the separator into
nested mappings.
.. note::
Keys not of type `str` are not supported and will raise errors.
:param mapping: the mapping to process
:param separator: th... | [
"def",
"_split_keys",
"(",
"mapping",
",",
"separator",
"=",
"'.'",
",",
"colliding",
"=",
"None",
")",
":",
"result",
"=",
"{",
"}",
"for",
"key",
",",
"value",
"in",
"mapping",
".",
"items",
"(",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"... | Recursively walks *mapping* to split keys that contain the separator into
nested mappings.
.. note::
Keys not of type `str` are not supported and will raise errors.
:param mapping: the mapping to process
:param separator: the character (sequence) to use as the separator between
keys
... | [
"Recursively",
"walks",
"*",
"mapping",
"*",
"to",
"split",
"keys",
"that",
"contain",
"the",
"separator",
"into",
"nested",
"mappings",
"."
] | train | https://github.com/HolmesNL/confidence/blob/e14d2d8769a01fa55676716f7a2f22714c2616d3/confidence/utils.py#L49-L91 |
tipsi/tipsi_tools | tipsi_tools/monitoring.py | log_mon_value | def log_mon_value(name, value=1, **kwargs):
"""
simplest monitoring function to be aggregated with sum
"""
message = '{} => {}'.format(name, value)
log_mon.info({'metric_name': name, 'value': value, 'message': message, **kwargs}) | python | def log_mon_value(name, value=1, **kwargs):
"""
simplest monitoring function to be aggregated with sum
"""
message = '{} => {}'.format(name, value)
log_mon.info({'metric_name': name, 'value': value, 'message': message, **kwargs}) | [
"def",
"log_mon_value",
"(",
"name",
",",
"value",
"=",
"1",
",",
"*",
"*",
"kwargs",
")",
":",
"message",
"=",
"'{} => {}'",
".",
"format",
"(",
"name",
",",
"value",
")",
"log_mon",
".",
"info",
"(",
"{",
"'metric_name'",
":",
"name",
",",
"'value'... | simplest monitoring function to be aggregated with sum | [
"simplest",
"monitoring",
"function",
"to",
"be",
"aggregated",
"with",
"sum"
] | train | https://github.com/tipsi/tipsi_tools/blob/1aba960c9890ceef2fb5e215b98b1646056ee58e/tipsi_tools/monitoring.py#L11-L16 |
alfredodeza/notario | notario/store.py | create_store | def create_store():
"""
A helper for setting the _proxy and slapping the store
object for us.
:return: A thread-local storage as a dictionary
"""
new_storage = _proxy('store')
_state.store = type('store', (object,), {})
new_storage.store = dict()
return new_storage.store | python | def create_store():
"""
A helper for setting the _proxy and slapping the store
object for us.
:return: A thread-local storage as a dictionary
"""
new_storage = _proxy('store')
_state.store = type('store', (object,), {})
new_storage.store = dict()
return new_storage.store | [
"def",
"create_store",
"(",
")",
":",
"new_storage",
"=",
"_proxy",
"(",
"'store'",
")",
"_state",
".",
"store",
"=",
"type",
"(",
"'store'",
",",
"(",
"object",
",",
")",
",",
"{",
"}",
")",
"new_storage",
".",
"store",
"=",
"dict",
"(",
")",
"ret... | A helper for setting the _proxy and slapping the store
object for us.
:return: A thread-local storage as a dictionary | [
"A",
"helper",
"for",
"setting",
"the",
"_proxy",
"and",
"slapping",
"the",
"store",
"object",
"for",
"us",
"."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/store.py#L25-L35 |
tipsi/tipsi_tools | tipsi_tools/drf/__init__.py | use_form | def use_form(form_class, request=None, **top_kwargs):
"""
Validate request (query_params or request body with args from url) with serializer and pass
validated data dict to the view function instead of request object.
"""
def validated_form(request, **kwargs):
# import ipdb; ipdb.set_trace(... | python | def use_form(form_class, request=None, **top_kwargs):
"""
Validate request (query_params or request body with args from url) with serializer and pass
validated data dict to the view function instead of request object.
"""
def validated_form(request, **kwargs):
# import ipdb; ipdb.set_trace(... | [
"def",
"use_form",
"(",
"form_class",
",",
"request",
"=",
"None",
",",
"*",
"*",
"top_kwargs",
")",
":",
"def",
"validated_form",
"(",
"request",
",",
"*",
"*",
"kwargs",
")",
":",
"# import ipdb; ipdb.set_trace()",
"data",
"=",
"request",
".",
"query_param... | Validate request (query_params or request body with args from url) with serializer and pass
validated data dict to the view function instead of request object. | [
"Validate",
"request",
"(",
"query_params",
"or",
"request",
"body",
"with",
"args",
"from",
"url",
")",
"with",
"serializer",
"and",
"pass",
"validated",
"data",
"dict",
"to",
"the",
"view",
"function",
"instead",
"of",
"request",
"object",
"."
] | train | https://github.com/tipsi/tipsi_tools/blob/1aba960c9890ceef2fb5e215b98b1646056ee58e/tipsi_tools/drf/__init__.py#L5-L49 |
craigahobbs/chisel | src/chisel/request.py | request | def request(request_callback=None, **kwargs):
"""
Chisel request decorator
"""
if request_callback is None:
return lambda fn: request(fn, **kwargs)
else:
return Request(request_callback, **kwargs).decorate_module(request_callback) | python | def request(request_callback=None, **kwargs):
"""
Chisel request decorator
"""
if request_callback is None:
return lambda fn: request(fn, **kwargs)
else:
return Request(request_callback, **kwargs).decorate_module(request_callback) | [
"def",
"request",
"(",
"request_callback",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"request_callback",
"is",
"None",
":",
"return",
"lambda",
"fn",
":",
"request",
"(",
"fn",
",",
"*",
"*",
"kwargs",
")",
"else",
":",
"return",
"Request",... | Chisel request decorator | [
"Chisel",
"request",
"decorator"
] | train | https://github.com/craigahobbs/chisel/blob/d306a9eae2ff757647c6ca1c933bc944efa5c326/src/chisel/request.py#L13-L21 |
bierschenk/ode | examples/example_2d_orbit.py | dx_orbit_sys | def dx_orbit_sys(t, X):
'''X = [
m1x, m1y,
m2x, m2y,
m3x, m3y,
m4x, m4y,
m1vx, m1vy,
m2vx, m2vy,
m3vx, m3vy,
m4vx, m4vy
]
'''
(m1x, m1y,
m2x, m2y,
m3x, m3y,
m4x, m4y,
m1vx, m1vy,
m2vx, m2vy,
m3vx, m3vy,
m4vx, m4vy) = X
m_moon1 = 7.34... | python | def dx_orbit_sys(t, X):
'''X = [
m1x, m1y,
m2x, m2y,
m3x, m3y,
m4x, m4y,
m1vx, m1vy,
m2vx, m2vy,
m3vx, m3vy,
m4vx, m4vy
]
'''
(m1x, m1y,
m2x, m2y,
m3x, m3y,
m4x, m4y,
m1vx, m1vy,
m2vx, m2vy,
m3vx, m3vy,
m4vx, m4vy) = X
m_moon1 = 7.34... | [
"def",
"dx_orbit_sys",
"(",
"t",
",",
"X",
")",
":",
"(",
"m1x",
",",
"m1y",
",",
"m2x",
",",
"m2y",
",",
"m3x",
",",
"m3y",
",",
"m4x",
",",
"m4y",
",",
"m1vx",
",",
"m1vy",
",",
"m2vx",
",",
"m2vy",
",",
"m3vx",
",",
"m3vy",
",",
"m4vx",
... | X = [
m1x, m1y,
m2x, m2y,
m3x, m3y,
m4x, m4y,
m1vx, m1vy,
m2vx, m2vy,
m3vx, m3vy,
m4vx, m4vy
] | [
"X",
"=",
"[",
"m1x",
"m1y",
"m2x",
"m2y",
"m3x",
"m3y",
"m4x",
"m4y",
"m1vx",
"m1vy",
"m2vx",
"m2vy",
"m3vx",
"m3vy",
"m4vx",
"m4vy",
"]"
] | train | https://github.com/bierschenk/ode/blob/01fb714874926f0988a4bb250d2a0c8a2429e4f0/examples/example_2d_orbit.py#L7-L76 |
Parsely/redis-fluster | fluster/penalty_box.py | PenaltyBox.add | def add(self, client):
"""Add a client to the penalty box."""
if client.pool_id in self._client_ids:
log.info("%r is already in the penalty box. Ignoring.", client)
return
release = time.time() + self._min_wait
heapq.heappush(self._clients, (release, (client, self... | python | def add(self, client):
"""Add a client to the penalty box."""
if client.pool_id in self._client_ids:
log.info("%r is already in the penalty box. Ignoring.", client)
return
release = time.time() + self._min_wait
heapq.heappush(self._clients, (release, (client, self... | [
"def",
"add",
"(",
"self",
",",
"client",
")",
":",
"if",
"client",
".",
"pool_id",
"in",
"self",
".",
"_client_ids",
":",
"log",
".",
"info",
"(",
"\"%r is already in the penalty box. Ignoring.\"",
",",
"client",
")",
"return",
"release",
"=",
"time",
".",
... | Add a client to the penalty box. | [
"Add",
"a",
"client",
"to",
"the",
"penalty",
"box",
"."
] | train | https://github.com/Parsely/redis-fluster/blob/9fb3ccdc3e0b24906520cac1e933a775e8dfbd99/fluster/penalty_box.py#L21-L28 |
Parsely/redis-fluster | fluster/penalty_box.py | PenaltyBox.get | def get(self):
"""Get any clients ready to be used.
:returns: Iterable of redis clients
"""
now = time.time()
while self._clients and self._clients[0][0] < now:
_, (client, last_wait) = heapq.heappop(self._clients)
connect_start = time.time()
... | python | def get(self):
"""Get any clients ready to be used.
:returns: Iterable of redis clients
"""
now = time.time()
while self._clients and self._clients[0][0] < now:
_, (client, last_wait) = heapq.heappop(self._clients)
connect_start = time.time()
... | [
"def",
"get",
"(",
"self",
")",
":",
"now",
"=",
"time",
".",
"time",
"(",
")",
"while",
"self",
".",
"_clients",
"and",
"self",
".",
"_clients",
"[",
"0",
"]",
"[",
"0",
"]",
"<",
"now",
":",
"_",
",",
"(",
"client",
",",
"last_wait",
")",
"... | Get any clients ready to be used.
:returns: Iterable of redis clients | [
"Get",
"any",
"clients",
"ready",
"to",
"be",
"used",
"."
] | train | https://github.com/Parsely/redis-fluster/blob/9fb3ccdc3e0b24906520cac1e933a775e8dfbd99/fluster/penalty_box.py#L30-L52 |
alfredodeza/notario | notario/validators/types.py | string | def string(_object):
"""
Validates a given input is of type string.
Example usage::
data = {'a' : 21}
schema = (string, 21)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argum... | python | def string(_object):
"""
Validates a given input is of type string.
Example usage::
data = {'a' : 21}
schema = (string, 21)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argum... | [
"def",
"string",
"(",
"_object",
")",
":",
"if",
"is_callable",
"(",
"_object",
")",
":",
"_validator",
"=",
"_object",
"@",
"wraps",
"(",
"_validator",
")",
"def",
"decorated",
"(",
"value",
")",
":",
"ensure",
"(",
"isinstance",
"(",
"value",
",",
"b... | Validates a given input is of type string.
Example usage::
data = {'a' : 21}
schema = (string, 21)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argument is a callable, the decorating... | [
"Validates",
"a",
"given",
"input",
"is",
"of",
"type",
"string",
"."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L10-L34 |
alfredodeza/notario | notario/validators/types.py | boolean | def boolean(_object):
"""
Validates a given input is of type boolean.
Example usage::
data = {'a' : True}
schema = ('a', boolean)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the... | python | def boolean(_object):
"""
Validates a given input is of type boolean.
Example usage::
data = {'a' : True}
schema = ('a', boolean)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the... | [
"def",
"boolean",
"(",
"_object",
")",
":",
"if",
"is_callable",
"(",
"_object",
")",
":",
"_validator",
"=",
"_object",
"@",
"wraps",
"(",
"_validator",
")",
"def",
"decorated",
"(",
"value",
")",
":",
"ensure",
"(",
"isinstance",
"(",
"value",
",",
"... | Validates a given input is of type boolean.
Example usage::
data = {'a' : True}
schema = ('a', boolean)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argument is a callable, the decor... | [
"Validates",
"a",
"given",
"input",
"is",
"of",
"type",
"boolean",
"."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L37-L62 |
alfredodeza/notario | notario/validators/types.py | dictionary | def dictionary(_object, *args):
"""
Validates a given input is of type dictionary.
Example usage::
data = {'a' : {'b': 1}}
schema = ('a', dictionary)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. n... | python | def dictionary(_object, *args):
"""
Validates a given input is of type dictionary.
Example usage::
data = {'a' : {'b': 1}}
schema = ('a', dictionary)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. n... | [
"def",
"dictionary",
"(",
"_object",
",",
"*",
"args",
")",
":",
"error_msg",
"=",
"'not of type dictionary'",
"if",
"is_callable",
"(",
"_object",
")",
":",
"_validator",
"=",
"_object",
"@",
"wraps",
"(",
"_validator",
")",
"def",
"decorated",
"(",
"value"... | Validates a given input is of type dictionary.
Example usage::
data = {'a' : {'b': 1}}
schema = ('a', dictionary)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argument is a callable,... | [
"Validates",
"a",
"given",
"input",
"is",
"of",
"type",
"dictionary",
"."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L66-L98 |
alfredodeza/notario | notario/validators/types.py | array | def array(_object):
"""
Validates a given input is of type list.
Example usage::
data = {'a' : [1,2]}
schema = ('a', array)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argum... | python | def array(_object):
"""
Validates a given input is of type list.
Example usage::
data = {'a' : [1,2]}
schema = ('a', array)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argum... | [
"def",
"array",
"(",
"_object",
")",
":",
"if",
"is_callable",
"(",
"_object",
")",
":",
"_validator",
"=",
"_object",
"@",
"wraps",
"(",
"_validator",
")",
"def",
"decorated",
"(",
"value",
")",
":",
"ensure",
"(",
"isinstance",
"(",
"value",
",",
"li... | Validates a given input is of type list.
Example usage::
data = {'a' : [1,2]}
schema = ('a', array)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argument is a callable, the decoratin... | [
"Validates",
"a",
"given",
"input",
"is",
"of",
"type",
"list",
"."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L101-L126 |
alfredodeza/notario | notario/validators/types.py | integer | def integer(_object):
"""
Validates a given input is of type int..
Example usage::
data = {'a' : 21}
schema = ('a', integer)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argu... | python | def integer(_object):
"""
Validates a given input is of type int..
Example usage::
data = {'a' : 21}
schema = ('a', integer)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argu... | [
"def",
"integer",
"(",
"_object",
")",
":",
"if",
"is_callable",
"(",
"_object",
")",
":",
"_validator",
"=",
"_object",
"@",
"wraps",
"(",
"_validator",
")",
"def",
"decorated",
"(",
"value",
")",
":",
"ensure",
"(",
"isinstance",
"(",
"value",
",",
"... | Validates a given input is of type int..
Example usage::
data = {'a' : 21}
schema = ('a', integer)
You can also use this as a decorator, as a way to check for the
input before it even hits a validator you may be writing.
.. note::
If the argument is a callable, the decorating... | [
"Validates",
"a",
"given",
"input",
"is",
"of",
"type",
"int",
".."
] | train | https://github.com/alfredodeza/notario/blob/d5dc2edfcb75d9291ced3f2551f368c35dd31475/notario/validators/types.py#L129-L153 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.constant | def constant(cls,
value: Value,
dtype: tf.DType = tf.float32) -> 'TensorFluent':
'''Returns a constant `value` TensorFluent with given `dtype`.
Args:
value: The constant value.
dtype: The output's data type.
Returns:
A constant Tensor... | python | def constant(cls,
value: Value,
dtype: tf.DType = tf.float32) -> 'TensorFluent':
'''Returns a constant `value` TensorFluent with given `dtype`.
Args:
value: The constant value.
dtype: The output's data type.
Returns:
A constant Tensor... | [
"def",
"constant",
"(",
"cls",
",",
"value",
":",
"Value",
",",
"dtype",
":",
"tf",
".",
"DType",
"=",
"tf",
".",
"float32",
")",
"->",
"'TensorFluent'",
":",
"t",
"=",
"tf",
".",
"constant",
"(",
"value",
",",
"dtype",
"=",
"dtype",
")",
"scope",
... | Returns a constant `value` TensorFluent with given `dtype`.
Args:
value: The constant value.
dtype: The output's data type.
Returns:
A constant TensorFluent. | [
"Returns",
"a",
"constant",
"value",
"TensorFluent",
"with",
"given",
"dtype",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L67-L82 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.Bernoulli | def Bernoulli(cls,
mean: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Bernoulli sampling op with given mean parameter.
Args:
mean: The mean parameter of the Bernoulli distribution.
bat... | python | def Bernoulli(cls,
mean: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Bernoulli sampling op with given mean parameter.
Args:
mean: The mean parameter of the Bernoulli distribution.
bat... | [
"def",
"Bernoulli",
"(",
"cls",
",",
"mean",
":",
"'TensorFluent'",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Distribution",
",",
"'TensorFluent'",
"]",
":",
"probs",
"=",
"mean",
".",
"tensor",
"dist",
... | Returns a TensorFluent for the Bernoulli sampling op with given mean parameter.
Args:
mean: The mean parameter of the Bernoulli distribution.
batch_size: The size of the batch (optional).
Returns:
The Bernoulli distribution and a TensorFluent sample drawn from the d... | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"Bernoulli",
"sampling",
"op",
"with",
"given",
"mean",
"parameter",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L85-L106 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.Uniform | def Uniform(cls,
low: 'TensorFluent', high: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Uniform sampling op with given low and high parameters.
Args:
low: The low parameter of the Uniform... | python | def Uniform(cls,
low: 'TensorFluent', high: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Uniform sampling op with given low and high parameters.
Args:
low: The low parameter of the Uniform... | [
"def",
"Uniform",
"(",
"cls",
",",
"low",
":",
"'TensorFluent'",
",",
"high",
":",
"'TensorFluent'",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Distribution",
",",
"'TensorFluent'",
"]",
":",
"if",
"low",
... | Returns a TensorFluent for the Uniform sampling op with given low and high parameters.
Args:
low: The low parameter of the Uniform distribution.
high: The high parameter of the Uniform distribution.
batch_size: The size of the batch (optional).
Returns:
... | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"Uniform",
"sampling",
"op",
"with",
"given",
"low",
"and",
"high",
"parameters",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L109-L135 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.Normal | def Normal(cls,
mean: 'TensorFluent', variance: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Normal sampling op with given mean and variance.
Args:
mean: The mean parameter of the Normal d... | python | def Normal(cls,
mean: 'TensorFluent', variance: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Normal sampling op with given mean and variance.
Args:
mean: The mean parameter of the Normal d... | [
"def",
"Normal",
"(",
"cls",
",",
"mean",
":",
"'TensorFluent'",
",",
"variance",
":",
"'TensorFluent'",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Distribution",
",",
"'TensorFluent'",
"]",
":",
"if",
"me... | Returns a TensorFluent for the Normal sampling op with given mean and variance.
Args:
mean: The mean parameter of the Normal distribution.
variance: The variance parameter of the Normal distribution.
batch_size: The size of the batch (optional).
Returns:
... | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"Normal",
"sampling",
"op",
"with",
"given",
"mean",
"and",
"variance",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L138-L166 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.Gamma | def Gamma(cls,
shape: 'TensorFluent',
scale: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters.
Args:
shape: The shape parame... | python | def Gamma(cls,
shape: 'TensorFluent',
scale: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters.
Args:
shape: The shape parame... | [
"def",
"Gamma",
"(",
"cls",
",",
"shape",
":",
"'TensorFluent'",
",",
"scale",
":",
"'TensorFluent'",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Distribution",
",",
"'TensorFluent'",
"]",
":",
"if",
"shape... | Returns a TensorFluent for the Gamma sampling op with given shape and scale parameters.
Args:
shape: The shape parameter of the Gamma distribution.
scale: The scale parameter of the Gamma distribution.
batch_size: The size of the batch (optional).
Returns:
... | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"Gamma",
"sampling",
"op",
"with",
"given",
"shape",
"and",
"scale",
"parameters",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L200-L229 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.Exponential | def Exponential(cls,
mean: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Exponential sampling op with given mean parameter.
Args:
mean: The mean parameter of the Exponential distribution.
... | python | def Exponential(cls,
mean: 'TensorFluent',
batch_size: Optional[int] = None) -> Tuple[Distribution, 'TensorFluent']:
'''Returns a TensorFluent for the Exponential sampling op with given mean parameter.
Args:
mean: The mean parameter of the Exponential distribution.
... | [
"def",
"Exponential",
"(",
"cls",
",",
"mean",
":",
"'TensorFluent'",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
")",
"->",
"Tuple",
"[",
"Distribution",
",",
"'TensorFluent'",
"]",
":",
"rate",
"=",
"1",
"/",
"mean",
".",
"tenso... | Returns a TensorFluent for the Exponential sampling op with given mean parameter.
Args:
mean: The mean parameter of the Exponential distribution.
batch_size: The size of the batch (optional).
Returns:
The Exponential distribution and a TensorFluent sample drawn from... | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"Exponential",
"sampling",
"op",
"with",
"given",
"mean",
"parameter",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L232-L253 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.stop_gradient | def stop_gradient(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a copy of the input fluent with stop_gradient at tensor level.
Args:
x: The input fluent.
Returns:
A TensorFluent that stops backpropagation of gradient computations.
'''
scope = x.s... | python | def stop_gradient(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a copy of the input fluent with stop_gradient at tensor level.
Args:
x: The input fluent.
Returns:
A TensorFluent that stops backpropagation of gradient computations.
'''
scope = x.s... | [
"def",
"stop_gradient",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"scope",
"=",
"x",
".",
"scope",
".",
"as_list",
"(",
")",
"batch",
"=",
"x",
".",
"batch",
"return",
"TensorFluent",
"(",
"tf",
".",
"stop_gradient",
... | Returns a copy of the input fluent with stop_gradient at tensor level.
Args:
x: The input fluent.
Returns:
A TensorFluent that stops backpropagation of gradient computations. | [
"Returns",
"a",
"copy",
"of",
"the",
"input",
"fluent",
"with",
"stop_gradient",
"at",
"tensor",
"level",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L256-L267 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.stop_batch_gradient | def stop_batch_gradient(cls, x: 'TensorFluent', stop_batch: tf.Tensor) -> 'TensorFluent':
'''Returns a copy of the inputs fluent with stop_gradient applied at batch level.
Args:
x: The input fluent.
stop_batch: A boolean tf.Tensor with shape=(batch_size, ...)
Returns:
... | python | def stop_batch_gradient(cls, x: 'TensorFluent', stop_batch: tf.Tensor) -> 'TensorFluent':
'''Returns a copy of the inputs fluent with stop_gradient applied at batch level.
Args:
x: The input fluent.
stop_batch: A boolean tf.Tensor with shape=(batch_size, ...)
Returns:
... | [
"def",
"stop_batch_gradient",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
",",
"stop_batch",
":",
"tf",
".",
"Tensor",
")",
"->",
"'TensorFluent'",
":",
"scope",
"=",
"x",
".",
"scope",
".",
"as_list",
"(",
")",
"batch",
"=",
"x",
".",
"batch",
"tenso... | Returns a copy of the inputs fluent with stop_gradient applied at batch level.
Args:
x: The input fluent.
stop_batch: A boolean tf.Tensor with shape=(batch_size, ...)
Returns:
A TensorFluent that conditionally stops backpropagation of gradient computations. | [
"Returns",
"a",
"copy",
"of",
"the",
"inputs",
"fluent",
"with",
"stop_gradient",
"applied",
"at",
"batch",
"level",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L270-L283 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.abs | def abs(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the abs function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the abs function.
'''
return cls._unary_op(x, tf.abs, tf.float32) | python | def abs(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the abs function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the abs function.
'''
return cls._unary_op(x, tf.abs, tf.float32) | [
"def",
"abs",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"abs",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the abs function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the abs function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"abs",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L286-L295 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.exp | def exp(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the exp function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the exp function.
'''
return cls._unary_op(x, tf.exp, tf.float32) | python | def exp(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the exp function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the exp function.
'''
return cls._unary_op(x, tf.exp, tf.float32) | [
"def",
"exp",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"exp",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the exp function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the exp function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"exp",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L298-L307 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.log | def log(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the log function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the log function.
'''
return cls._unary_op(x, tf.log, tf.float32) | python | def log(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the log function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the log function.
'''
return cls._unary_op(x, tf.log, tf.float32) | [
"def",
"log",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"log",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the log function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the log function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"log",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L310-L319 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.sqrt | def sqrt(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the sqrt function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sqrt function.
'''
return cls._unary_op(x, tf.sqrt, tf.float32) | python | def sqrt(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the sqrt function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sqrt function.
'''
return cls._unary_op(x, tf.sqrt, tf.float32) | [
"def",
"sqrt",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"sqrt",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the sqrt function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sqrt function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"sqrt",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L322-L331 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.cos | def cos(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the cos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the cos function.
'''
return cls._unary_op(x, tf.cos, tf.float32) | python | def cos(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the cos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the cos function.
'''
return cls._unary_op(x, tf.cos, tf.float32) | [
"def",
"cos",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"cos",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the cos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the cos function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"cos",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L334-L343 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.sin | def sin(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the sin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sin function.
'''
return cls._unary_op(x, tf.sin, tf.float32) | python | def sin(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the sin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sin function.
'''
return cls._unary_op(x, tf.sin, tf.float32) | [
"def",
"sin",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"sin",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the sin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the sin function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"sin",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L346-L355 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.tan | def tan(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the tan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the tan function.
'''
return cls._unary_op(x, tf.tan, tf.float32) | python | def tan(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the tan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the tan function.
'''
return cls._unary_op(x, tf.tan, tf.float32) | [
"def",
"tan",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"tan",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the tan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the tan function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"tan",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L358-L367 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.acos | def acos(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arccos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arccos function.
'''
return cls._unary_op(x, tf.acos, tf.float32) | python | def acos(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arccos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arccos function.
'''
return cls._unary_op(x, tf.acos, tf.float32) | [
"def",
"acos",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"acos",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the arccos function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arccos function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"arccos",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L370-L379 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.asin | def asin(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arcsin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arcsin function.
'''
return cls._unary_op(x, tf.asin, tf.float32) | python | def asin(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arcsin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arcsin function.
'''
return cls._unary_op(x, tf.asin, tf.float32) | [
"def",
"asin",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"asin",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the arcsin function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arcsin function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"arcsin",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L382-L391 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.atan | def atan(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arctan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arctan function.
'''
return cls._unary_op(x, tf.atan2, tf.float32) | python | def atan(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the arctan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arctan function.
'''
return cls._unary_op(x, tf.atan2, tf.float32) | [
"def",
"atan",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"atan2",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the arctan function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the arctan function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"arctan",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L394-L403 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.round | def round(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the round function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the round function.
'''
return cls._unary_op(x, tf.round, tf.float32) | python | def round(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the round function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the round function.
'''
return cls._unary_op(x, tf.round, tf.float32) | [
"def",
"round",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"round",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the round function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the round function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"round",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L406-L415 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.ceil | def ceil(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the ceil function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the ceil function.
'''
return cls._unary_op(x, tf.ceil, tf.float32) | python | def ceil(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the ceil function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the ceil function.
'''
return cls._unary_op(x, tf.ceil, tf.float32) | [
"def",
"ceil",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"ceil",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the ceil function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the ceil function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"ceil",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L418-L427 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.floor | def floor(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the floor function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the floor function.
'''
return cls._unary_op(x, tf.floor, tf.float32) | python | def floor(cls, x: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the floor function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the floor function.
'''
return cls._unary_op(x, tf.floor, tf.float32) | [
"def",
"floor",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_unary_op",
"(",
"x",
",",
"tf",
".",
"floor",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the floor function.
Args:
x: The input fluent.
Returns:
A TensorFluent wrapping the floor function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"floor",
"function",
"."
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L430-L439 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.pow | def pow(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the pow function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the pow function.
'''
return... | python | def pow(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the pow function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the pow function.
'''
return... | [
"def",
"pow",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
",",
"y",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_binary_op",
"(",
"x",
",",
"y",
",",
"tf",
".",
"pow",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the pow function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the pow function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"pow",
"function",
".",
"TensorFluent"
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L442-L452 |
thiagopbueno/rddl2tf | rddl2tf/fluent.py | TensorFluent.max | def max(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the maximum function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the maximum function.
'''
... | python | def max(cls, x: 'TensorFluent', y: 'TensorFluent') -> 'TensorFluent':
'''Returns a TensorFluent for the maximum function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the maximum function.
'''
... | [
"def",
"max",
"(",
"cls",
",",
"x",
":",
"'TensorFluent'",
",",
"y",
":",
"'TensorFluent'",
")",
"->",
"'TensorFluent'",
":",
"return",
"cls",
".",
"_binary_op",
"(",
"x",
",",
"y",
",",
"tf",
".",
"maximum",
",",
"tf",
".",
"float32",
")"
] | Returns a TensorFluent for the maximum function.TensorFluent
Args:
x: The first operand.
y: The second operand.
Returns:
A TensorFluent wrapping the maximum function. | [
"Returns",
"a",
"TensorFluent",
"for",
"the",
"maximum",
"function",
".",
"TensorFluent"
] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/fluent.py#L455-L465 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.