commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
10
3.52k
new_contents
stringlengths
21
3.18k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.32k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
fuzzy_diff
stringlengths
17
3.24k
ff800f11b948808e4574ec3a893ed4e259707dcf
stubs/python2-urllib2/run.py
stubs/python2-urllib2/run.py
import sys import ssl import urllib2 if len(sys.argv) < 3 or len(sys.argv) > 4: exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0]) host = sys.argv[1] port = sys.argv[2] cafile = sys.argv[3] if len(sys.argv) > 3 else None try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) except ssl.CertificateError: print("REJECT") except urllib2.URLError as exc: if not isinstance(exc.reason, ssl.SSLError): raise print("REJECT") else: print("ACCEPT")
import sys import ssl import urllib2 if len(sys.argv) < 3 or len(sys.argv) > 4: exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0]) host = sys.argv[1] port = sys.argv[2] cafile = sys.argv[3] if len(sys.argv) > 3 else None try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) except getattr(ssl, "CertificateError", ()): print("REJECT") except urllib2.URLError as exc: if not isinstance(exc.reason, ssl.SSLError): raise print("REJECT") else: print("ACCEPT")
Make python2-urllib2 compatible with more Python 2.7 versions
Make python2-urllib2 compatible with more Python 2.7 versions Try to catch ssl.CertificateError only if CertificateError is defined. Otherwise bail out by effectively doing a dummy "catch ():".
Python
mit
ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls
import sys import ssl import urllib2 if len(sys.argv) < 3 or len(sys.argv) > 4: exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0]) host = sys.argv[1] port = sys.argv[2] cafile = sys.argv[3] if len(sys.argv) > 3 else None try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) - except ssl.CertificateError: + except getattr(ssl, "CertificateError", ()): print("REJECT") except urllib2.URLError as exc: if not isinstance(exc.reason, ssl.SSLError): raise print("REJECT") else: print("ACCEPT")
Make python2-urllib2 compatible with more Python 2.7 versions
## Code Before: import sys import ssl import urllib2 if len(sys.argv) < 3 or len(sys.argv) > 4: exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0]) host = sys.argv[1] port = sys.argv[2] cafile = sys.argv[3] if len(sys.argv) > 3 else None try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) except ssl.CertificateError: print("REJECT") except urllib2.URLError as exc: if not isinstance(exc.reason, ssl.SSLError): raise print("REJECT") else: print("ACCEPT") ## Instruction: Make python2-urllib2 compatible with more Python 2.7 versions ## Code After: import sys import ssl import urllib2 if len(sys.argv) < 3 or len(sys.argv) > 4: exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0]) host = sys.argv[1] port = sys.argv[2] cafile = sys.argv[3] if len(sys.argv) > 3 else None try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) except getattr(ssl, "CertificateError", ()): print("REJECT") except urllib2.URLError as exc: if not isinstance(exc.reason, ssl.SSLError): raise print("REJECT") else: print("ACCEPT")
# ... existing code ... try: urllib2.urlopen("https://" + host + ":" + port, cafile=cafile) except getattr(ssl, "CertificateError", ()): print("REJECT") except urllib2.URLError as exc: # ... rest of the code ...
2f67880e777c9efa5192f5c34ce5fc7d71fc0f08
partner_communication_switzerland/wizards/end_contract_wizard.py
partner_communication_switzerland/wizards/end_contract_wizard.py
from odoo import models, fields, api class EndContractWizard(models.TransientModel): _inherit = 'end.contract.wizard' generate_communication = fields.Boolean( 'Create depart communication') @api.multi def end_contract(self): self.ensure_one() child = self.child_id if self.generate_communication: exit_config = self.env.ref( 'partner_communication_switzerland.' 'lifecycle_child_unplanned_exit') self.contract_id.with_context( default_object_ids=child.id, default_auto_send=False).send_communication(exit_config) return super(EndContractWizard, self).end_contract()
from odoo import models, fields, api class EndContractWizard(models.TransientModel): _inherit = 'end.contract.wizard' generate_communication = fields.Boolean( 'Create depart communication') @api.multi def end_contract(self): self.ensure_one() if self.generate_communication: exit_config = self.env.ref( 'partner_communication_switzerland.' 'lifecycle_child_unplanned_exit') self.contract_id.with_context( default_object_ids=self.contract_id.id, default_auto_send=False).send_communication(exit_config) return super(EndContractWizard, self).end_contract()
FIX end contract depart letter generation
FIX end contract depart letter generation
Python
agpl-3.0
eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland,ecino/compassion-switzerland,ecino/compassion-switzerland,ecino/compassion-switzerland,eicher31/compassion-switzerland,eicher31/compassion-switzerland
from odoo import models, fields, api class EndContractWizard(models.TransientModel): _inherit = 'end.contract.wizard' generate_communication = fields.Boolean( 'Create depart communication') @api.multi def end_contract(self): self.ensure_one() - child = self.child_id - if self.generate_communication: exit_config = self.env.ref( 'partner_communication_switzerland.' 'lifecycle_child_unplanned_exit') self.contract_id.with_context( - default_object_ids=child.id, + default_object_ids=self.contract_id.id, default_auto_send=False).send_communication(exit_config) return super(EndContractWizard, self).end_contract()
FIX end contract depart letter generation
## Code Before: from odoo import models, fields, api class EndContractWizard(models.TransientModel): _inherit = 'end.contract.wizard' generate_communication = fields.Boolean( 'Create depart communication') @api.multi def end_contract(self): self.ensure_one() child = self.child_id if self.generate_communication: exit_config = self.env.ref( 'partner_communication_switzerland.' 'lifecycle_child_unplanned_exit') self.contract_id.with_context( default_object_ids=child.id, default_auto_send=False).send_communication(exit_config) return super(EndContractWizard, self).end_contract() ## Instruction: FIX end contract depart letter generation ## Code After: from odoo import models, fields, api class EndContractWizard(models.TransientModel): _inherit = 'end.contract.wizard' generate_communication = fields.Boolean( 'Create depart communication') @api.multi def end_contract(self): self.ensure_one() if self.generate_communication: exit_config = self.env.ref( 'partner_communication_switzerland.' 'lifecycle_child_unplanned_exit') self.contract_id.with_context( default_object_ids=self.contract_id.id, default_auto_send=False).send_communication(exit_config) return super(EndContractWizard, self).end_contract()
// ... existing code ... def end_contract(self): self.ensure_one() if self.generate_communication: exit_config = self.env.ref( // ... modified code ... 'lifecycle_child_unplanned_exit') self.contract_id.with_context( default_object_ids=self.contract_id.id, default_auto_send=False).send_communication(exit_config) // ... rest of the code ...
477364a4d2895fc79af2a57ace35ededf0281911
mistral/db/sqlalchemy/migration/alembic_migrations/versions/003_cron_trigger_constraints.py
mistral/db/sqlalchemy/migration/alembic_migrations/versions/003_cron_trigger_constraints.py
# revision identifiers, used by Alembic. revision = '003' down_revision = '002' from alembic import op import sqlalchemy as sa def upgrade(): op.add_column( 'cron_triggers_v2', sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) op.drop_index('workflow_input_hash', table_name='cron_triggers_v2') op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2') op.create_unique_constraint( None, 'cron_triggers_v2', [ 'workflow_input_hash', 'workflow_name', 'pattern', 'project_id', 'workflow_params_hash', 'remaining_executions', 'first_execution_time' ] )
# revision identifiers, used by Alembic. revision = '003' down_revision = '002' from alembic import op import sqlalchemy as sa def upgrade(): op.add_column( 'cron_triggers_v2', sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) op.create_unique_constraint( None, 'cron_triggers_v2', [ 'workflow_input_hash', 'workflow_name', 'pattern', 'project_id', 'workflow_params_hash', 'remaining_executions', 'first_execution_time' ] )
Fix database upgrade from a new database
Fix database upgrade from a new database This fixes the problem where running "mistral-db-manage upgrade heads" on a new database result in error with workflow_input_hash index does not exist. Change-Id: I560b2b78d11cd3fd4ae9c8606e4336e87b22ef27 Closes-Bug: #1519929
Python
apache-2.0
openstack/mistral,StackStorm/mistral,openstack/mistral,StackStorm/mistral
# revision identifiers, used by Alembic. revision = '003' down_revision = '002' from alembic import op import sqlalchemy as sa def upgrade(): op.add_column( 'cron_triggers_v2', sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) + - op.drop_index('workflow_input_hash', table_name='cron_triggers_v2') - op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2') op.create_unique_constraint( None, 'cron_triggers_v2', [ 'workflow_input_hash', 'workflow_name', 'pattern', 'project_id', 'workflow_params_hash', 'remaining_executions', 'first_execution_time' ] )
Fix database upgrade from a new database
## Code Before: # revision identifiers, used by Alembic. revision = '003' down_revision = '002' from alembic import op import sqlalchemy as sa def upgrade(): op.add_column( 'cron_triggers_v2', sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) op.drop_index('workflow_input_hash', table_name='cron_triggers_v2') op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2') op.create_unique_constraint( None, 'cron_triggers_v2', [ 'workflow_input_hash', 'workflow_name', 'pattern', 'project_id', 'workflow_params_hash', 'remaining_executions', 'first_execution_time' ] ) ## Instruction: Fix database upgrade from a new database ## Code After: # revision identifiers, used by Alembic. revision = '003' down_revision = '002' from alembic import op import sqlalchemy as sa def upgrade(): op.add_column( 'cron_triggers_v2', sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) op.create_unique_constraint( None, 'cron_triggers_v2', [ 'workflow_input_hash', 'workflow_name', 'pattern', 'project_id', 'workflow_params_hash', 'remaining_executions', 'first_execution_time' ] )
... sa.Column('first_execution_time', sa.DateTime(), nullable=True) ) op.create_unique_constraint( None, ...
dd89173cc177f7130eca426eb4fa5737ec59c91d
test/vpp_mac.py
test/vpp_mac.py
from util import mactobinary class VppMacAddress(): def __init__(self, addr): self.address = addr def encode(self): return { 'bytes': self.bytes } @property def bytes(self): return mactobinary(self.address) @property def address(self): return self.addr.address def __str__(self): return self.address def __eq__(self, other): if isinstance(other, self.__class__): return self.address == other.addres elif hasattr(other, "bytes"): # vl_api_mac_addres_t return self.bytes == other.bytes else: raise Exception("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) return False
from util import mactobinary class VppMacAddress(): def __init__(self, addr): self.address = addr def encode(self): return { 'bytes': self.bytes } @property def bytes(self): return mactobinary(self.address) @property def address(self): return self.address @address.setter def address(self, value): self.address = value def __str__(self): return self.address def __eq__(self, other): if isinstance(other, self.__class__): return self.address == other.address elif hasattr(other, "bytes"): # vl_api_mac_addres_t return self.bytes == other.bytes else: raise TypeError("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) return False
Fix L2BD arp termination Test Case
Fix L2BD arp termination Test Case ============================================================================== L2BD arp termination Test Case ============================================================================== 12:02:21,850 Couldn't stat : /tmp/vpp-unittest-TestL2bdArpTerm-_h44qo/stats.sock L2BD arp term - add 5 hosts, verify arp responses OK L2BD arp term - delete 3 hosts, verify arp responses OK L2BD arp term - recreate BD1, readd 3 hosts, verify arp responses OK L2BD arp term - 2 IP4 addrs per host OK L2BD arp term - create and update 10 IP4-mac pairs OK L2BD arp/ND term - hosts with both ip4/ip6 OK L2BD ND term - Add and Del hosts, verify ND replies OK L2BD ND term - Add and update IP+mac, verify ND replies OK L2BD arp term - send garps, verify arp event reports OK L2BD arp term - send duplicate garps, verify suppression OK L2BD arp term - disable ip4 arp events,send garps, verify no events OK L2BD ND term - send NS packets verify reports OK L2BD ND term - send duplicate ns, verify suppression OK L2BD ND term - disable ip4 arp events,send ns, verify no events OK ============================================================================== TEST RESULTS: Scheduled tests: 14 Executed tests: 14 Passed tests: 14 ============================================================================== Test run was successful Change-Id: I6bb1ced11b88080ffaa845d22b0bc471c4f91683 Signed-off-by: Paul Vinciguerra <[email protected]>
Python
apache-2.0
chrisy/vpp,vpp-dev/vpp,FDio/vpp,FDio/vpp,FDio/vpp,FDio/vpp,chrisy/vpp,FDio/vpp,chrisy/vpp,vpp-dev/vpp,chrisy/vpp,vpp-dev/vpp,vpp-dev/vpp,vpp-dev/vpp,FDio/vpp,chrisy/vpp,chrisy/vpp,chrisy/vpp,vpp-dev/vpp,vpp-dev/vpp,FDio/vpp,chrisy/vpp,FDio/vpp
from util import mactobinary class VppMacAddress(): def __init__(self, addr): self.address = addr def encode(self): return { 'bytes': self.bytes } @property def bytes(self): return mactobinary(self.address) @property def address(self): - return self.addr.address + return self.address + + @address.setter + def address(self, value): + self.address = value def __str__(self): return self.address def __eq__(self, other): if isinstance(other, self.__class__): - return self.address == other.addres + return self.address == other.address elif hasattr(other, "bytes"): # vl_api_mac_addres_t return self.bytes == other.bytes else: - raise Exception("Comparing VppMacAddress:%s" + raise TypeError("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) return False
Fix L2BD arp termination Test Case
## Code Before: from util import mactobinary class VppMacAddress(): def __init__(self, addr): self.address = addr def encode(self): return { 'bytes': self.bytes } @property def bytes(self): return mactobinary(self.address) @property def address(self): return self.addr.address def __str__(self): return self.address def __eq__(self, other): if isinstance(other, self.__class__): return self.address == other.addres elif hasattr(other, "bytes"): # vl_api_mac_addres_t return self.bytes == other.bytes else: raise Exception("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) return False ## Instruction: Fix L2BD arp termination Test Case ## Code After: from util import mactobinary class VppMacAddress(): def __init__(self, addr): self.address = addr def encode(self): return { 'bytes': self.bytes } @property def bytes(self): return mactobinary(self.address) @property def address(self): return self.address @address.setter def address(self, value): self.address = value def __str__(self): return self.address def __eq__(self, other): if isinstance(other, self.__class__): return self.address == other.address elif hasattr(other, "bytes"): # vl_api_mac_addres_t return self.bytes == other.bytes else: raise TypeError("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) return False
// ... existing code ... @property def address(self): return self.address @address.setter def address(self, value): self.address = value def __str__(self): // ... modified code ... def __eq__(self, other): if isinstance(other, self.__class__): return self.address == other.address elif hasattr(other, "bytes"): # vl_api_mac_addres_t ... return self.bytes == other.bytes else: raise TypeError("Comparing VppMacAddress:%s" "with unknown type: %s" % (self, other)) // ... rest of the code ...
b9156a522410bf39de8653bce22bb2cb56e435a0
parktain/main.py
parktain/main.py
from os.path import abspath, dirname, join from gendo import Gendo HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') bot = Gendo.config_from_yaml(config_path) @bot.listen_for('morning') def morning(user, message): # make sure message is "morning" and doesn't just contain it. if message.strip() == "morning": return "mornin' @{user.username}" def main(): bot.run() if __name__ == '__main__': main()
from os.path import abspath, dirname, join import re # 3rd party library from gendo import Gendo class Parktain(Gendo): """Overridden to add simple additional functionality.""" @property def id(self): """Get id of the bot.""" if not hasattr(self, '_id',): self._id = self.client.server.login_data['self']['id'] return self._id @property def username(self): """Get username of the bot.""" if not hasattr(self, '_username',): self._username = self.client.server.username return self.username HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') bot = Parktain.config_from_yaml(config_path) def is_mention(f): """Decorator to check if bot is mentioned.""" def wrapped(name, message): BOT_ID_RE = re.compile('<@{}>'.format(bot.id)) mention = BOT_ID_RE.search(message) is not None if mention: return f(name, message) return wrapped @bot.listen_for('morning') def morning(user, message): # make sure message is "morning" and doesn't just contain it. if message.strip() == "morning": return "mornin' @{user.username}" @bot.listen_for('where do you live') @is_mention def source_code(user, message): repo_url = 'https://github.com/punchagan/parktain' message = 'Well, I live in your hearts...\nYou can change me from here {}, though.' return message.format(repo_url) def main(): bot.run() if __name__ == '__main__': main()
Add 'where do you live' answer.
Add 'where do you live' answer.
Python
bsd-3-clause
punchagan/parktain,punchagan/parktain,punchagan/parktain
+ from os.path import abspath, dirname, join + import re - from os.path import abspath, dirname, join + # 3rd party library from gendo import Gendo + + + class Parktain(Gendo): + """Overridden to add simple additional functionality.""" + + @property + def id(self): + """Get id of the bot.""" + + if not hasattr(self, '_id',): + self._id = self.client.server.login_data['self']['id'] + return self._id + + @property + def username(self): + """Get username of the bot.""" + + if not hasattr(self, '_username',): + self._username = self.client.server.username + return self.username + + HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') - bot = Gendo.config_from_yaml(config_path) + bot = Parktain.config_from_yaml(config_path) + + def is_mention(f): + """Decorator to check if bot is mentioned.""" + + def wrapped(name, message): + BOT_ID_RE = re.compile('<@{}>'.format(bot.id)) + mention = BOT_ID_RE.search(message) is not None + if mention: + return f(name, message) + + return wrapped @bot.listen_for('morning') def morning(user, message): # make sure message is "morning" and doesn't just contain it. if message.strip() == "morning": return "mornin' @{user.username}" + + @bot.listen_for('where do you live') + @is_mention + def source_code(user, message): + repo_url = 'https://github.com/punchagan/parktain' + message = 'Well, I live in your hearts...\nYou can change me from here {}, though.' + return message.format(repo_url) + + def main(): bot.run() if __name__ == '__main__': main()
Add 'where do you live' answer.
## Code Before: from os.path import abspath, dirname, join from gendo import Gendo HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') bot = Gendo.config_from_yaml(config_path) @bot.listen_for('morning') def morning(user, message): # make sure message is "morning" and doesn't just contain it. if message.strip() == "morning": return "mornin' @{user.username}" def main(): bot.run() if __name__ == '__main__': main() ## Instruction: Add 'where do you live' answer. ## Code After: from os.path import abspath, dirname, join import re # 3rd party library from gendo import Gendo class Parktain(Gendo): """Overridden to add simple additional functionality.""" @property def id(self): """Get id of the bot.""" if not hasattr(self, '_id',): self._id = self.client.server.login_data['self']['id'] return self._id @property def username(self): """Get username of the bot.""" if not hasattr(self, '_username',): self._username = self.client.server.username return self.username HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') bot = Parktain.config_from_yaml(config_path) def is_mention(f): """Decorator to check if bot is mentioned.""" def wrapped(name, message): BOT_ID_RE = re.compile('<@{}>'.format(bot.id)) mention = BOT_ID_RE.search(message) is not None if mention: return f(name, message) return wrapped @bot.listen_for('morning') def morning(user, message): # make sure message is "morning" and doesn't just contain it. if message.strip() == "morning": return "mornin' @{user.username}" @bot.listen_for('where do you live') @is_mention def source_code(user, message): repo_url = 'https://github.com/punchagan/parktain' message = 'Well, I live in your hearts...\nYou can change me from here {}, though.' return message.format(repo_url) def main(): bot.run() if __name__ == '__main__': main()
... from os.path import abspath, dirname, join import re # 3rd party library from gendo import Gendo class Parktain(Gendo): """Overridden to add simple additional functionality.""" @property def id(self): """Get id of the bot.""" if not hasattr(self, '_id',): self._id = self.client.server.login_data['self']['id'] return self._id @property def username(self): """Get username of the bot.""" if not hasattr(self, '_username',): self._username = self.client.server.username return self.username HERE = dirname(abspath(__file__)) config_path = join(HERE, 'config.yaml') bot = Parktain.config_from_yaml(config_path) def is_mention(f): """Decorator to check if bot is mentioned.""" def wrapped(name, message): BOT_ID_RE = re.compile('<@{}>'.format(bot.id)) mention = BOT_ID_RE.search(message) is not None if mention: return f(name, message) return wrapped ... return "mornin' @{user.username}" @bot.listen_for('where do you live') @is_mention def source_code(user, message): repo_url = 'https://github.com/punchagan/parktain' message = 'Well, I live in your hearts...\nYou can change me from here {}, though.' return message.format(repo_url) def main(): bot.run() ...
5d9fa1838ffe7ffedb59453a0eca520b5f8d5849
pyscf/ci/__init__.py
pyscf/ci/__init__.py
from pyscf.ci.cisd import CISD
from pyscf.ci import cisd def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None): from pyscf import scf if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)): raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version') return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
Revert accidental changes to ci
Revert accidental changes to ci
Python
apache-2.0
gkc1000/pyscf,gkc1000/pyscf,gkc1000/pyscf,sunqm/pyscf,sunqm/pyscf,sunqm/pyscf,gkc1000/pyscf,gkc1000/pyscf,sunqm/pyscf
- from pyscf.ci.cisd import CISD + from pyscf.ci import cisd + + def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None): + from pyscf import scf + if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)): + raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version') + return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
Revert accidental changes to ci
## Code Before: from pyscf.ci.cisd import CISD ## Instruction: Revert accidental changes to ci ## Code After: from pyscf.ci import cisd def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None): from pyscf import scf if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)): raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version') return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
// ... existing code ... from pyscf.ci import cisd def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None): from pyscf import scf if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)): raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version') return cisd.CISD(mf, frozen, mo_coeff, mo_occ) // ... rest of the code ...
0f7853c3568791f0e93ece57d2fc750dbc93b963
starlette/concurrency.py
starlette/concurrency.py
import asyncio import functools import typing from typing import Any, AsyncGenerator, Iterator try: import contextvars # Python 3.7+ only. except ImportError: # pragma: no cover contextvars = None # type: ignore async def run_in_threadpool( func: typing.Callable, *args: typing.Any, **kwargs: typing.Any ) -> typing.Any: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover # Ensure we run in the same context child = functools.partial(func, *args, **kwargs) context = contextvars.copy_context() func = context.run args = (child,) elif kwargs: # pragma: no cover # loop.run_in_executor doesn't accept 'kwargs', so bind them in here func = functools.partial(func, **kwargs) return await loop.run_in_executor(None, func, *args) class _StopIteration(Exception): pass def _next(iterator: Iterator) -> Any: # We can't raise `StopIteration` from within the threadpool iterator # and catch it outside that context, so we coerce them into a different # exception type. try: return next(iterator) except StopIteration: raise _StopIteration async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator: while True: try: yield await run_in_threadpool(_next, iterator) except _StopIteration: break
import asyncio import functools import typing from typing import Any, AsyncGenerator, Iterator try: import contextvars # Python 3.7+ only. except ImportError: # pragma: no cover contextvars = None # type: ignore T = typing.TypeVar("T") async def run_in_threadpool( func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any ) -> T: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover # Ensure we run in the same context child = functools.partial(func, *args, **kwargs) context = contextvars.copy_context() func = context.run args = (child,) elif kwargs: # pragma: no cover # loop.run_in_executor doesn't accept 'kwargs', so bind them in here func = functools.partial(func, **kwargs) return await loop.run_in_executor(None, func, *args) class _StopIteration(Exception): pass def _next(iterator: Iterator) -> Any: # We can't raise `StopIteration` from within the threadpool iterator # and catch it outside that context, so we coerce them into a different # exception type. try: return next(iterator) except StopIteration: raise _StopIteration async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator: while True: try: yield await run_in_threadpool(_next, iterator) except _StopIteration: break
Add type hint for run_in_threadpool return type
Add type hint for run_in_threadpool return type
Python
bsd-3-clause
encode/starlette,encode/starlette
import asyncio import functools import typing from typing import Any, AsyncGenerator, Iterator try: import contextvars # Python 3.7+ only. except ImportError: # pragma: no cover contextvars = None # type: ignore + T = typing.TypeVar("T") + async def run_in_threadpool( - func: typing.Callable, *args: typing.Any, **kwargs: typing.Any + func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any - ) -> typing.Any: + ) -> T: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover # Ensure we run in the same context child = functools.partial(func, *args, **kwargs) context = contextvars.copy_context() func = context.run args = (child,) elif kwargs: # pragma: no cover # loop.run_in_executor doesn't accept 'kwargs', so bind them in here func = functools.partial(func, **kwargs) return await loop.run_in_executor(None, func, *args) class _StopIteration(Exception): pass def _next(iterator: Iterator) -> Any: # We can't raise `StopIteration` from within the threadpool iterator # and catch it outside that context, so we coerce them into a different # exception type. try: return next(iterator) except StopIteration: raise _StopIteration async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator: while True: try: yield await run_in_threadpool(_next, iterator) except _StopIteration: break
Add type hint for run_in_threadpool return type
## Code Before: import asyncio import functools import typing from typing import Any, AsyncGenerator, Iterator try: import contextvars # Python 3.7+ only. except ImportError: # pragma: no cover contextvars = None # type: ignore async def run_in_threadpool( func: typing.Callable, *args: typing.Any, **kwargs: typing.Any ) -> typing.Any: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover # Ensure we run in the same context child = functools.partial(func, *args, **kwargs) context = contextvars.copy_context() func = context.run args = (child,) elif kwargs: # pragma: no cover # loop.run_in_executor doesn't accept 'kwargs', so bind them in here func = functools.partial(func, **kwargs) return await loop.run_in_executor(None, func, *args) class _StopIteration(Exception): pass def _next(iterator: Iterator) -> Any: # We can't raise `StopIteration` from within the threadpool iterator # and catch it outside that context, so we coerce them into a different # exception type. try: return next(iterator) except StopIteration: raise _StopIteration async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator: while True: try: yield await run_in_threadpool(_next, iterator) except _StopIteration: break ## Instruction: Add type hint for run_in_threadpool return type ## Code After: import asyncio import functools import typing from typing import Any, AsyncGenerator, Iterator try: import contextvars # Python 3.7+ only. except ImportError: # pragma: no cover contextvars = None # type: ignore T = typing.TypeVar("T") async def run_in_threadpool( func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any ) -> T: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover # Ensure we run in the same context child = functools.partial(func, *args, **kwargs) context = contextvars.copy_context() func = context.run args = (child,) elif kwargs: # pragma: no cover # loop.run_in_executor doesn't accept 'kwargs', so bind them in here func = functools.partial(func, **kwargs) return await loop.run_in_executor(None, func, *args) class _StopIteration(Exception): pass def _next(iterator: Iterator) -> Any: # We can't raise `StopIteration` from within the threadpool iterator # and catch it outside that context, so we coerce them into a different # exception type. try: return next(iterator) except StopIteration: raise _StopIteration async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator: while True: try: yield await run_in_threadpool(_next, iterator) except _StopIteration: break
// ... existing code ... contextvars = None # type: ignore T = typing.TypeVar("T") async def run_in_threadpool( func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any ) -> T: loop = asyncio.get_event_loop() if contextvars is not None: # pragma: no cover // ... rest of the code ...
7d02bd555d7519d485d00e02136d26a6e4e7096e
nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py
nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py
from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
Drop FK before dropping instance_id column.
Drop FK before dropping instance_id column.
Python
apache-2.0
sacharya/nova,jianghuaw/nova,leilihh/novaha,eneabio/nova,vladikr/nova_drafts,KarimAllah/nova,sileht/deb-openstack-nova,Stavitsky/nova,DirectXMan12/nova-hacking,akash1808/nova_test_latest,raildo/nova,gspilio/nova,tangfeixiong/nova,jianghuaw/nova,Juniper/nova,JioCloud/nova,zhimin711/nova,usc-isi/nova,orbitfp7/nova,JianyuWang/nova,vmturbo/nova,sebrandon1/nova,jeffrey4l/nova,Francis-Liu/animated-broccoli,psiwczak/openstack,MountainWei/nova,tianweizhang/nova,yrobla/nova,maelnor/nova,whitepages/nova,maoy/zknova,joker946/nova,russellb/nova,iuliat/nova,qwefi/nova,rahulunair/nova,berrange/nova,sileht/deb-openstack-nova,mahak/nova,fnordahl/nova,sridevikoushik31/openstack,Metaswitch/calico-nova,gooddata/openstack-nova,sebrandon1/nova,redhat-openstack/nova,eayunstack/nova,mandeepdhami/nova,tealover/nova,eharney/nova,yrobla/nova,CEG-FYP-OpenStack/scheduler,TieWei/nova,maelnor/nova,TwinkleChawla/nova,KarimAllah/nova,cloudbau/nova,isyippee/nova,mikalstill/nova,hanlind/nova,mgagne/nova,badock/nova,qwefi/nova,paulmathews/nova,kimjaejoong/nova,spring-week-topos/nova-week,plumgrid/plumgrid-nova,alaski/nova,petrutlucian94/nova,thomasem/nova,barnsnake351/nova,cernops/nova,akash1808/nova,Triv90/Nova,yrobla/nova,watonyweng/nova,akash1808/nova_test_latest,NoBodyCam/TftpPxeBootBareMetal,Tehsmash/nova,Juniper/nova,iuliat/nova,orbitfp7/nova,alexandrucoman/vbox-nova-driver,aristanetworks/arista-ovs-nova,fnordahl/nova,cernops/nova,zaina/nova,projectcalico/calico-nova,russellb/nova,apporc/nova,j-carpentier/nova,shahar-stratoscale/nova,DirectXMan12/nova-hacking,tealover/nova,vmturbo/nova,rahulunair/nova,JianyuWang/nova,varunarya10/nova_test_latest,imsplitbit/nova,klmitch/nova,silenceli/nova,NewpTone/stacklab-nova,apporc/nova,devendermishrajio/nova_test_latest,dawnpower/nova,alvarolopez/nova,felixma/nova,saleemjaveds/https-github.com-openstack-nova,adelina-t/nova,angdraug/nova,mikalstill/nova,akash1808/nova,Yuriy-Leonov/nova,CiscoSystems/nova,klmitch/nova,watonyweng/nova,devoid/nova,bgxavier/nova,citrix-openstack-build/nova,psiwczak/openstack,nikesh-mahalka/nova,sridevikoushik31/nova,CiscoSystems/nova,joker946/nova,JioCloud/nova,salv-orlando/MyRepo,rrader/nova-docker-plugin,kimjaejoong/nova,rickerc/nova_audit,savi-dev/nova,sridevikoushik31/nova,hanlind/nova,DirectXMan12/nova-hacking,blueboxgroup/nova,JioCloud/nova_test_latest,eonpatapon/nova,luogangyi/bcec-nova,belmiromoreira/nova,fajoy/nova,rickerc/nova_audit,double12gzh/nova,sileht/deb-openstack-nova,cloudbase/nova,eayunstack/nova,NeCTAR-RC/nova,aristanetworks/arista-ovs-nova,CCI-MOC/nova,sridevikoushik31/openstack,silenceli/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,virtualopensystems/nova,Juniper/nova,devendermishrajio/nova,tudorvio/nova,edulramirez/nova,bgxavier/nova,cyx1231st/nova,shootstar/novatest,varunarya10/nova_test_latest,maheshp/novatest,cernops/nova,imsplitbit/nova,maheshp/novatest,russellb/nova,josephsuh/extra-specs,mahak/nova,mgagne/nova,plumgrid/plumgrid-nova,gspilio/nova,sridevikoushik31/nova,luogangyi/bcec-nova,NoBodyCam/TftpPxeBootBareMetal,alaski/nova,cloudbau/nova,CloudServer/nova,bigswitch/nova,houshengbo/nova_vmware_compute_driver,dawnpower/nova,rajalokan/nova,belmiromoreira/nova,virtualopensystems/nova,saleemjaveds/https-github.com-openstack-nova,bclau/nova,eonpatapon/nova,Juniper/nova,citrix-openstack-build/nova,j-carpentier/nova,sacharya/nova,zhimin711/nova,Yusuke1987/openstack_template,angdraug/nova,mmnelemane/nova,eneabio/nova,cloudbase/nova,klmitch/nova,vmturbo/nova,openstack/nova,zaina/nova,edulramirez/nova,eharney/nova,josephsuh/extra-specs,cloudbase/nova,shail2810/nova,jianghuaw/nova,Triv90/Nova,NeCTAR-RC/nova,viggates/nova,zzicewind/nova,LoHChina/nova,vmturbo/nova,spring-week-topos/nova-week,noironetworks/nova,rajalokan/nova,openstack/nova,berrange/nova,takeshineshiro/nova,eneabio/nova,cloudbase/nova-virtualbox,felixma/nova,fajoy/nova,whitepages/nova,usc-isi/extra-specs,psiwczak/openstack,ruslanloman/nova,isyippee/nova,ruslanloman/nova,petrutlucian94/nova_dev,dstroppa/openstack-smartos-nova-grizzly,shahar-stratoscale/nova,bclau/nova,josephsuh/extra-specs,SUSE-Cloud/nova,vladikr/nova_drafts,noironetworks/nova,fajoy/nova,ntt-sic/nova,maoy/zknova,Francis-Liu/animated-broccoli,BeyondTheClouds/nova,blueboxgroup/nova,LoHChina/nova,cloudbase/nova-virtualbox,Triv90/Nova,jianghuaw/nova,SUSE-Cloud/nova,leilihh/novaha,devoid/nova,salv-orlando/MyRepo,Yuriy-Leonov/nova,jeffrey4l/nova,NewpTone/stacklab-nova,tangfeixiong/nova,zzicewind/nova,houshengbo/nova_vmware_compute_driver,yosshy/nova,BeyondTheClouds/nova,sridevikoushik31/openstack,aristanetworks/arista-ovs-nova,maheshp/novatest,OpenAcademy-OpenStack/nova-scheduler,mandeepdhami/nova,phenoxim/nova,paulmathews/nova,usc-isi/nova,TwinkleChawla/nova,mikalstill/nova,Metaswitch/calico-nova,ntt-sic/nova,KarimAllah/nova,houshengbo/nova_vmware_compute_driver,projectcalico/calico-nova,CloudServer/nova,savi-dev/nova,usc-isi/extra-specs,tanglei528/nova,yatinkumbhare/openstack-nova,tianweizhang/nova,Stavitsky/nova,gooddata/openstack-nova,redhat-openstack/nova,Yusuke1987/openstack_template,sridevikoushik31/nova,yatinkumbhare/openstack-nova,tanglei528/nova,leilihh/nova,dstroppa/openstack-smartos-nova-grizzly,klmitch/nova,mahak/nova,sebrandon1/nova,bigswitch/nova,rajalokan/nova,rrader/nova-docker-plugin,leilihh/nova,raildo/nova,mmnelemane/nova,ewindisch/nova,dstroppa/openstack-smartos-nova-grizzly,gooddata/openstack-nova,Tehsmash/nova,JioCloud/nova_test_latest,petrutlucian94/nova_dev,ted-gould/nova,rahulunair/nova,dims/nova,badock/nova,yosshy/nova,MountainWei/nova,scripnichenko/nova,double12gzh/nova,gspilio/nova,OpenAcademy-OpenStack/nova-scheduler,nikesh-mahalka/nova,shootstar/novatest,savi-dev/nova,scripnichenko/nova,gooddata/openstack-nova,devendermishrajio/nova_test_latest,usc-isi/extra-specs,CCI-MOC/nova,takeshineshiro/nova,usc-isi/nova,NewpTone/stacklab-nova,dims/nova,adelina-t/nova,phenoxim/nova,shail2810/nova,cyx1231st/nova,alvarolopez/nova,alexandrucoman/vbox-nova-driver,barnsnake351/nova,openstack/nova,affo/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,affo/nova,maoy/zknova,thomasem/nova,ted-gould/nova,petrutlucian94/nova,viggates/nova,CEG-FYP-OpenStack/scheduler,TieWei/nova,salv-orlando/MyRepo,hanlind/nova,rajalokan/nova,NoBodyCam/TftpPxeBootBareMetal,ewindisch/nova,BeyondTheClouds/nova,devendermishrajio/nova,paulmathews/nova,tudorvio/nova
from sqlalchemy import Column, Integer, String, MetaData, Table + meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) + + if migrate_engine.name == "mysql": + migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ + "`migrations_ibfk_1`;") + migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
Drop FK before dropping instance_id column.
## Code Before: from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id) ## Instruction: Drop FK before dropping instance_id column. ## Code After: from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
# ... existing code ... from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # ... modified code ... migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() # ... rest of the code ...
23072e882edb6da55cb12ef0591a786235249670
ome/__main__.py
ome/__main__.py
import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
Use print_verbose for conditional printing.
Use print_verbose for conditional printing.
Python
mit
shaurz/ome,shaurz/ome
import sys from .command import command_args from .error import OmeError from .terminal import stderr + + def print_verbose(*args, **kwargs): + if command_args.verbose: + print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) - if command_args.verbose: - print('ome: using target {}'.format(target.name)) + print_verbose('ome: using target {}'.format(target.name)) - print('ome: using backend {} {}'.format(backend.name, backend.version)) + print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: - if command_args.verbose: - print('ome: compiling {}'.format(filename)) + print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
Use print_verbose for conditional printing.
## Code Before: import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main() ## Instruction: Use print_verbose for conditional printing. ## Code After: import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
// ... existing code ... from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): // ... modified code ... build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) // ... rest of the code ...
4b43a2f50740bbeab95f64137eb8993ed8ac4617
other/password_generator.py
other/password_generator.py
import string from random import * letters = string.ascii_letters digits = string.digits symbols = string.punctuation chars = letters + digits + symbols min_length = 8 max_length = 16 password = ''.join(choice(chars) for x in range(randint(min_length, max_length))) print('Password: %s' % password) print('[ If you are thinking of using this passsword, You better save it. ]')
import string import random letters = [letter for letter in string.ascii_letters] digits = [digit for digit in string.digits] symbols = [symbol for symbol in string.punctuation] chars = letters + digits + symbols random.shuffle(chars) min_length = 8 max_length = 16 password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length))) print('Password: ' + password) print('[ If you are thinking of using this passsword, You better save it. ]')
Add another randomness into the password generator
Add another randomness into the password generator Uses import random for namespace cleanliness Uses list instead of string for 'chars' variable in order to shuffle, increases randomness Instead of string formatting, uses string concatenation because (currently) it is simpler
Python
mit
TheAlgorithms/Python
import string - from random import * + import random - letters = string.ascii_letters - digits = string.digits - symbols = string.punctuation + letters = [letter for letter in string.ascii_letters] + digits = [digit for digit in string.digits] + symbols = [symbol for symbol in string.punctuation] chars = letters + digits + symbols + random.shuffle(chars) min_length = 8 max_length = 16 - password = ''.join(choice(chars) for x in range(randint(min_length, max_length))) + password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length))) - print('Password: %s' % password) + print('Password: ' + password) print('[ If you are thinking of using this passsword, You better save it. ]')
Add another randomness into the password generator
## Code Before: import string from random import * letters = string.ascii_letters digits = string.digits symbols = string.punctuation chars = letters + digits + symbols min_length = 8 max_length = 16 password = ''.join(choice(chars) for x in range(randint(min_length, max_length))) print('Password: %s' % password) print('[ If you are thinking of using this passsword, You better save it. ]') ## Instruction: Add another randomness into the password generator ## Code After: import string import random letters = [letter for letter in string.ascii_letters] digits = [digit for digit in string.digits] symbols = [symbol for symbol in string.punctuation] chars = letters + digits + symbols random.shuffle(chars) min_length = 8 max_length = 16 password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length))) print('Password: ' + password) print('[ If you are thinking of using this passsword, You better save it. ]')
... import string import random letters = [letter for letter in string.ascii_letters] digits = [digit for digit in string.digits] symbols = [symbol for symbol in string.punctuation] chars = letters + digits + symbols random.shuffle(chars) min_length = 8 max_length = 16 password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length))) print('Password: ' + password) print('[ If you are thinking of using this passsword, You better save it. ]') ...
ba98874be9370ec49c2c04e89d456f723b5d083c
monitoring/test/test_data/exceptions.py
monitoring/test/test_data/exceptions.py
from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions def data(TEST): TEST.exceptions = exceptions.data monitoring_exception = monascacli.ClientException TEST.exceptions.monitoring = exceptions.create_stubbed_exception( monitoring_exception)
try: from monascaclient.apiclient import exceptions as monascacli except ImportError: from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions def data(TEST): TEST.exceptions = exceptions.data monitoring_exception = monascacli.ClientException TEST.exceptions.monitoring = exceptions.create_stubbed_exception( monitoring_exception)
Adjust tests for python-monascaclient >= 1.3.0
Adjust tests for python-monascaclient >= 1.3.0 the exceptions module was moved out of the openstack.common namespace, so try to import the new location first and fall back to the old one if it doesn't exist. Change-Id: I3305775baaab15dca8d5e7e5cfc0932f94d4d153
Python
apache-2.0
openstack/monasca-ui,openstack/monasca-ui,openstack/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,openstack/monasca-ui
+ try: + from monascaclient.apiclient import exceptions as monascacli + except ImportError: + from monascaclient.openstack.common.apiclient import exceptions as monascacli - from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions def data(TEST): TEST.exceptions = exceptions.data monitoring_exception = monascacli.ClientException TEST.exceptions.monitoring = exceptions.create_stubbed_exception( monitoring_exception)
Adjust tests for python-monascaclient >= 1.3.0
## Code Before: from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions def data(TEST): TEST.exceptions = exceptions.data monitoring_exception = monascacli.ClientException TEST.exceptions.monitoring = exceptions.create_stubbed_exception( monitoring_exception) ## Instruction: Adjust tests for python-monascaclient >= 1.3.0 ## Code After: try: from monascaclient.apiclient import exceptions as monascacli except ImportError: from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions def data(TEST): TEST.exceptions = exceptions.data monitoring_exception = monascacli.ClientException TEST.exceptions.monitoring = exceptions.create_stubbed_exception( monitoring_exception)
# ... existing code ... try: from monascaclient.apiclient import exceptions as monascacli except ImportError: from monascaclient.openstack.common.apiclient import exceptions as monascacli from openstack_dashboard.test.test_data import exceptions # ... rest of the code ...
b2b1c2b8543cae37990262b2a811a9b0f26327da
arm/utils/locker.py
arm/utils/locker.py
from kvs import CacheKvs class Locker(object): """ locker for move the locker """ LOCKER_KEY = 'locker' EXPIRES = 5 # 5 sec def __init__(self, key=None): self.key = self.LOCKER_KEY if key: self.key += '.{}'.format(key) self.locker = CacheKvs(self.key) def lock(self): self.locker.set('locked', expires=self.EXPIRES, nx=True) def unlock(self): self.locker.delete() def is_lock(self): return self.locker.get() == 'locked' def on_lock(self, func): def wrapper(*args, **kwargs): if self.is_lock(): return self.lock() try: return func(*args, **kwargs) except Exception as e: raise e finally: self.unlock() return wrapper
from kvs import CacheKvs class Locker(object): """ locker for move the locker """ LOCKER_KEY = 'locker' EXPIRES = 5 # 5 sec def __init__(self, key=None): self.key = self.LOCKER_KEY if key: self.key += '.{}'.format(key) self.locker = CacheKvs(self.key) def lock(self): self.locker.set('locked', expires=self.EXPIRES, nx=True) def unlock(self): self.locker.delete() def is_lock(self): return self.locker.get() == 'locked' def on_lock(self, func): def wrapper(*args, **kwargs): if self.lock(): try: return func(*args, **kwargs) except Exception as e: raise e finally: self.unlock() return wrapper
Fix redis lock, use SETNX
Fix redis lock, use SETNX
Python
mit
mapler/tuesday,mapler/tuesday,mapler/tuesday
from kvs import CacheKvs class Locker(object): """ locker for move the locker """ LOCKER_KEY = 'locker' EXPIRES = 5 # 5 sec def __init__(self, key=None): self.key = self.LOCKER_KEY if key: self.key += '.{}'.format(key) self.locker = CacheKvs(self.key) def lock(self): self.locker.set('locked', expires=self.EXPIRES, nx=True) def unlock(self): self.locker.delete() def is_lock(self): return self.locker.get() == 'locked' def on_lock(self, func): def wrapper(*args, **kwargs): - if self.is_lock(): + if self.lock(): - return - self.lock() - try: + try: - return func(*args, **kwargs) + return func(*args, **kwargs) - except Exception as e: + except Exception as e: - raise e + raise e - finally: + finally: - self.unlock() + self.unlock() return wrapper
Fix redis lock, use SETNX
## Code Before: from kvs import CacheKvs class Locker(object): """ locker for move the locker """ LOCKER_KEY = 'locker' EXPIRES = 5 # 5 sec def __init__(self, key=None): self.key = self.LOCKER_KEY if key: self.key += '.{}'.format(key) self.locker = CacheKvs(self.key) def lock(self): self.locker.set('locked', expires=self.EXPIRES, nx=True) def unlock(self): self.locker.delete() def is_lock(self): return self.locker.get() == 'locked' def on_lock(self, func): def wrapper(*args, **kwargs): if self.is_lock(): return self.lock() try: return func(*args, **kwargs) except Exception as e: raise e finally: self.unlock() return wrapper ## Instruction: Fix redis lock, use SETNX ## Code After: from kvs import CacheKvs class Locker(object): """ locker for move the locker """ LOCKER_KEY = 'locker' EXPIRES = 5 # 5 sec def __init__(self, key=None): self.key = self.LOCKER_KEY if key: self.key += '.{}'.format(key) self.locker = CacheKvs(self.key) def lock(self): self.locker.set('locked', expires=self.EXPIRES, nx=True) def unlock(self): self.locker.delete() def is_lock(self): return self.locker.get() == 'locked' def on_lock(self, func): def wrapper(*args, **kwargs): if self.lock(): try: return func(*args, **kwargs) except Exception as e: raise e finally: self.unlock() return wrapper
// ... existing code ... def on_lock(self, func): def wrapper(*args, **kwargs): if self.lock(): try: return func(*args, **kwargs) except Exception as e: raise e finally: self.unlock() return wrapper // ... rest of the code ...
d1628356c7981748e2446c7b43d33d21cdef7e02
geoengine_partner/geo_partner.py
geoengine_partner/geo_partner.py
from openerp.osv import fields from base_geoengine import geo_model class ResPartner(geo_model.GeoModel): """Add geo_point to partner using a function filed""" _name = "res.partner" _inherit = "res.partner" _columns = { 'geo_point': fields.geo_point('Addresses coordinate') }
from openerp.osv import fields from openerp.addons.base_geoengine import geo_model class ResPartner(geo_model.GeoModel): """Add geo_point to partner using a function filed""" _name = "res.partner" _inherit = "res.partner" _columns = { 'geo_point': fields.geo_point('Addresses coordinate') }
Use absolute imports on opnerp.addons
[FIX] Use absolute imports on opnerp.addons
Python
agpl-3.0
OCA/geospatial,OCA/geospatial,OCA/geospatial
from openerp.osv import fields - from base_geoengine import geo_model + from openerp.addons.base_geoengine import geo_model class ResPartner(geo_model.GeoModel): """Add geo_point to partner using a function filed""" _name = "res.partner" _inherit = "res.partner" _columns = { 'geo_point': fields.geo_point('Addresses coordinate') }
Use absolute imports on opnerp.addons
## Code Before: from openerp.osv import fields from base_geoengine import geo_model class ResPartner(geo_model.GeoModel): """Add geo_point to partner using a function filed""" _name = "res.partner" _inherit = "res.partner" _columns = { 'geo_point': fields.geo_point('Addresses coordinate') } ## Instruction: Use absolute imports on opnerp.addons ## Code After: from openerp.osv import fields from openerp.addons.base_geoengine import geo_model class ResPartner(geo_model.GeoModel): """Add geo_point to partner using a function filed""" _name = "res.partner" _inherit = "res.partner" _columns = { 'geo_point': fields.geo_point('Addresses coordinate') }
... from openerp.osv import fields from openerp.addons.base_geoengine import geo_model ...
8b07dde78e753f6dce663481a68856024ed2fc49
plutokore/__init__.py
plutokore/__init__.py
from .environments.makino import MakinoProfile from .environments.king import KingProfile from .jet import AstroJet from . import luminosity from . import plotting from . import simulations from . import helpers from . import io __all__ = [ 'environments', 'luminosity', 'plotting', 'simulations', 'jet', 'helpers', 'io', ]
from .environments.makino import MakinoProfile from .environments.king import KingProfile from .jet import AstroJet from . import luminosity from . import plotting from . import simulations from . import helpers from . import io from . import configuration __all__ = [ 'environments', 'luminosity', 'plotting', 'simulations', 'jet', 'helpers', 'io', 'configuration', ]
Add configuration module to package exports
Add configuration module to package exports
Python
mit
opcon/plutokore,opcon/plutokore
from .environments.makino import MakinoProfile from .environments.king import KingProfile from .jet import AstroJet from . import luminosity from . import plotting from . import simulations from . import helpers from . import io + from . import configuration __all__ = [ 'environments', 'luminosity', 'plotting', 'simulations', 'jet', 'helpers', 'io', + 'configuration', ]
Add configuration module to package exports
## Code Before: from .environments.makino import MakinoProfile from .environments.king import KingProfile from .jet import AstroJet from . import luminosity from . import plotting from . import simulations from . import helpers from . import io __all__ = [ 'environments', 'luminosity', 'plotting', 'simulations', 'jet', 'helpers', 'io', ] ## Instruction: Add configuration module to package exports ## Code After: from .environments.makino import MakinoProfile from .environments.king import KingProfile from .jet import AstroJet from . import luminosity from . import plotting from . import simulations from . import helpers from . import io from . import configuration __all__ = [ 'environments', 'luminosity', 'plotting', 'simulations', 'jet', 'helpers', 'io', 'configuration', ]
# ... existing code ... from . import helpers from . import io from . import configuration __all__ = [ # ... modified code ... 'helpers', 'io', 'configuration', ] # ... rest of the code ...
01036133ed749d96a74bafb6b3f8670c06c63a84
1selfOpenDashboardCommand.py
1selfOpenDashboardCommand.py
import sublime, sublime_plugin, webbrowser QD_URL = "https://app.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): def run(self,edit): SETTINGS = {} SETTINGS_FILE = "1self.sublime-settings" SETTINGS = sublime.load_settings(SETTINGS_FILE) stream_id = SETTINGS.get("streamId") read_token = SETTINGS.get("readToken") VERSION = SETTINGS.get("VERSION") qd_url = QD_URL url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url)
import sublime, sublime_plugin, webbrowser QD_URL = "http://www.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): def run(self,edit): SETTINGS = {} SETTINGS_FILE = "1self.sublime-settings" SETTINGS = sublime.load_settings(SETTINGS_FILE) stream_id = SETTINGS.get("streamId") read_token = SETTINGS.get("readToken") VERSION = SETTINGS.get("VERSION") qd_url = QD_URL url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url)
Change landing URLs to website
Change landing URLs to website
Python
apache-2.0
1self/sublime-text-plugin,1self/sublime-text-plugin,1self/sublime-text-plugin
import sublime, sublime_plugin, webbrowser - QD_URL = "https://app.1self.co" + QD_URL = "http://www.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): def run(self,edit): SETTINGS = {} SETTINGS_FILE = "1self.sublime-settings" SETTINGS = sublime.load_settings(SETTINGS_FILE) stream_id = SETTINGS.get("streamId") read_token = SETTINGS.get("readToken") VERSION = SETTINGS.get("VERSION") qd_url = QD_URL - url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() + url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url)
Change landing URLs to website
## Code Before: import sublime, sublime_plugin, webbrowser QD_URL = "https://app.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): def run(self,edit): SETTINGS = {} SETTINGS_FILE = "1self.sublime-settings" SETTINGS = sublime.load_settings(SETTINGS_FILE) stream_id = SETTINGS.get("streamId") read_token = SETTINGS.get("readToken") VERSION = SETTINGS.get("VERSION") qd_url = QD_URL url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url) ## Instruction: Change landing URLs to website ## Code After: import sublime, sublime_plugin, webbrowser QD_URL = "http://www.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): def run(self,edit): SETTINGS = {} SETTINGS_FILE = "1self.sublime-settings" SETTINGS = sublime.load_settings(SETTINGS_FILE) stream_id = SETTINGS.get("streamId") read_token = SETTINGS.get("readToken") VERSION = SETTINGS.get("VERSION") qd_url = QD_URL url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url)
... QD_URL = "http://www.1self.co" class GoTo1selfDashboardCommand(sublime_plugin.TextCommand): ... qd_url = QD_URL url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals() print(url) webbrowser.open_new_tab(url) ...
668a5240c29047d86fe9451f3078bb163bea0db9
skan/__init__.py
skan/__init__.py
from .csr import skeleton_to_csgraph, branch_statistics, summarise __all__ = ['skeleton_to_csgraph', 'branch_statistics', 'summarise']
from .csr import skeleton_to_csgraph, branch_statistics, summarise __version__ = '0.1-dev' __all__ = ['skeleton_to_csgraph', 'branch_statistics', 'summarise']
Add version info to package init
Add version info to package init
Python
bsd-3-clause
jni/skan
from .csr import skeleton_to_csgraph, branch_statistics, summarise + + __version__ = '0.1-dev' __all__ = ['skeleton_to_csgraph', 'branch_statistics', 'summarise'] +
Add version info to package init
## Code Before: from .csr import skeleton_to_csgraph, branch_statistics, summarise __all__ = ['skeleton_to_csgraph', 'branch_statistics', 'summarise'] ## Instruction: Add version info to package init ## Code After: from .csr import skeleton_to_csgraph, branch_statistics, summarise __version__ = '0.1-dev' __all__ = ['skeleton_to_csgraph', 'branch_statistics', 'summarise']
# ... existing code ... from .csr import skeleton_to_csgraph, branch_statistics, summarise __version__ = '0.1-dev' __all__ = ['skeleton_to_csgraph', # ... rest of the code ...
b5f980b700707ecc611746f93b1f62650c76c451
pgcrypto_fields/aggregates.py
pgcrypto_fields/aggregates.py
from django.db import models class Decrypt(models.Aggregate): """`Decrypt` creates an alias for `DecryptFunctionSQL`. `alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`. `self.lookup` is defined in `models.Aggregate.__init__`. """ def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" from pgcrypto_fields.sql import aggregates klass = getattr(aggregates, self.name) aggregate = klass( col, source=source, is_summary=is_summary, **self.extra ) query.aggregates[alias] = aggregate class PGPPub(Decrypt): """PGP public key based aggregation.""" name = 'PGPPub' class PGPSym(Decrypt): """PGP symmetric key based aggregation.""" name = 'PGPSym'
from django.db import models from pgcrypto_fields.sql import aggregates class Decrypt(models.Aggregate): """`Decrypt` creates an alias for `DecryptFunctionSQL`. `alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`. `self.lookup` is defined in `models.Aggregate.__init__`. """ def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" klass = getattr(aggregates, self.name) aggregate = klass( col, source=source, is_summary=is_summary, **self.extra ) query.aggregates[alias] = aggregate class PGPPub(Decrypt): """PGP public key based aggregation.""" name = 'PGPPub' class PGPSym(Decrypt): """PGP symmetric key based aggregation.""" name = 'PGPSym'
Move import to top of the file
Move import to top of the file
Python
bsd-2-clause
incuna/django-pgcrypto-fields,atdsaa/django-pgcrypto-fields
from django.db import models + + from pgcrypto_fields.sql import aggregates class Decrypt(models.Aggregate): """`Decrypt` creates an alias for `DecryptFunctionSQL`. `alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`. `self.lookup` is defined in `models.Aggregate.__init__`. """ def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" - from pgcrypto_fields.sql import aggregates - klass = getattr(aggregates, self.name) aggregate = klass( col, source=source, is_summary=is_summary, **self.extra ) query.aggregates[alias] = aggregate class PGPPub(Decrypt): """PGP public key based aggregation.""" name = 'PGPPub' class PGPSym(Decrypt): """PGP symmetric key based aggregation.""" name = 'PGPSym'
Move import to top of the file
## Code Before: from django.db import models class Decrypt(models.Aggregate): """`Decrypt` creates an alias for `DecryptFunctionSQL`. `alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`. `self.lookup` is defined in `models.Aggregate.__init__`. """ def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" from pgcrypto_fields.sql import aggregates klass = getattr(aggregates, self.name) aggregate = klass( col, source=source, is_summary=is_summary, **self.extra ) query.aggregates[alias] = aggregate class PGPPub(Decrypt): """PGP public key based aggregation.""" name = 'PGPPub' class PGPSym(Decrypt): """PGP symmetric key based aggregation.""" name = 'PGPSym' ## Instruction: Move import to top of the file ## Code After: from django.db import models from pgcrypto_fields.sql import aggregates class Decrypt(models.Aggregate): """`Decrypt` creates an alias for `DecryptFunctionSQL`. `alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`. `self.lookup` is defined in `models.Aggregate.__init__`. """ def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" klass = getattr(aggregates, self.name) aggregate = klass( col, source=source, is_summary=is_summary, **self.extra ) query.aggregates[alias] = aggregate class PGPPub(Decrypt): """PGP public key based aggregation.""" name = 'PGPPub' class PGPSym(Decrypt): """PGP symmetric key based aggregation.""" name = 'PGPSym'
# ... existing code ... from django.db import models from pgcrypto_fields.sql import aggregates # ... modified code ... def add_to_query(self, query, alias, col, source, is_summary): """Add the aggregate to the query.""" klass = getattr(aggregates, self.name) aggregate = klass( # ... rest of the code ...
7fc4e7382665cf9eac4d19efcf9641ad57271e87
organizer/models.py
organizer/models.py
from django.db import models # Model Field Reference # https://docs.djangoproject.com/en/1.8/ref/models/fields/ class Tag(models.Model): name = models.CharField( max_length=31, unique=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') class Meta: ordering = ['name'] def __str__(self): return self.name class Startup(models.Model): name = models.CharField( max_length=31, db_index=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') description = models.TextField() founded_date = models.DateField( 'date founded') contact = models.EmailField() website = models.URLField(max_length=255) tags = models.ManyToManyField(Tag) def __str__(self): return self.name class NewsLink(models.Model): title = models.CharField(max_length=63) pub_date = models.DateField('date published') link = models.URLField(max_length=255) startup = models.ForeignKey(Startup) def __str__(self): return "{}: {}".format( self.startup, self.title)
from django.db import models # Model Field Reference # https://docs.djangoproject.com/en/1.8/ref/models/fields/ class Tag(models.Model): name = models.CharField( max_length=31, unique=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') class Meta: ordering = ['name'] def __str__(self): return self.name class Startup(models.Model): name = models.CharField( max_length=31, db_index=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') description = models.TextField() founded_date = models.DateField( 'date founded') contact = models.EmailField() website = models.URLField(max_length=255) tags = models.ManyToManyField(Tag) class Meta: ordering = ['name'] get_latest_by = 'founded_date' def __str__(self): return self.name class NewsLink(models.Model): title = models.CharField(max_length=63) pub_date = models.DateField('date published') link = models.URLField(max_length=255) startup = models.ForeignKey(Startup) def __str__(self): return "{}: {}".format( self.startup, self.title)
Declare Meta class in Startup model.
Ch03: Declare Meta class in Startup model. [skip ci]
Python
bsd-2-clause
jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8
from django.db import models # Model Field Reference # https://docs.djangoproject.com/en/1.8/ref/models/fields/ class Tag(models.Model): name = models.CharField( max_length=31, unique=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') class Meta: ordering = ['name'] def __str__(self): return self.name class Startup(models.Model): name = models.CharField( max_length=31, db_index=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') description = models.TextField() founded_date = models.DateField( 'date founded') contact = models.EmailField() website = models.URLField(max_length=255) tags = models.ManyToManyField(Tag) + class Meta: + ordering = ['name'] + get_latest_by = 'founded_date' + def __str__(self): return self.name class NewsLink(models.Model): title = models.CharField(max_length=63) pub_date = models.DateField('date published') link = models.URLField(max_length=255) startup = models.ForeignKey(Startup) def __str__(self): return "{}: {}".format( self.startup, self.title)
Declare Meta class in Startup model.
## Code Before: from django.db import models # Model Field Reference # https://docs.djangoproject.com/en/1.8/ref/models/fields/ class Tag(models.Model): name = models.CharField( max_length=31, unique=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') class Meta: ordering = ['name'] def __str__(self): return self.name class Startup(models.Model): name = models.CharField( max_length=31, db_index=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') description = models.TextField() founded_date = models.DateField( 'date founded') contact = models.EmailField() website = models.URLField(max_length=255) tags = models.ManyToManyField(Tag) def __str__(self): return self.name class NewsLink(models.Model): title = models.CharField(max_length=63) pub_date = models.DateField('date published') link = models.URLField(max_length=255) startup = models.ForeignKey(Startup) def __str__(self): return "{}: {}".format( self.startup, self.title) ## Instruction: Declare Meta class in Startup model. ## Code After: from django.db import models # Model Field Reference # https://docs.djangoproject.com/en/1.8/ref/models/fields/ class Tag(models.Model): name = models.CharField( max_length=31, unique=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') class Meta: ordering = ['name'] def __str__(self): return self.name class Startup(models.Model): name = models.CharField( max_length=31, db_index=True) slug = models.SlugField( max_length=31, unique=True, help_text='A label for URL config.') description = models.TextField() founded_date = models.DateField( 'date founded') contact = models.EmailField() website = models.URLField(max_length=255) tags = models.ManyToManyField(Tag) class Meta: ordering = ['name'] get_latest_by = 'founded_date' def __str__(self): return self.name class NewsLink(models.Model): title = models.CharField(max_length=63) pub_date = models.DateField('date published') link = models.URLField(max_length=255) startup = models.ForeignKey(Startup) def __str__(self): return "{}: {}".format( self.startup, self.title)
# ... existing code ... tags = models.ManyToManyField(Tag) class Meta: ordering = ['name'] get_latest_by = 'founded_date' def __str__(self): return self.name # ... rest of the code ...
d44010acc32fcb78570cd34478d0f4e8f1cfa979
utility/dbproc.py
utility/dbproc.py
from discord.ext import commands from utils import * from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from member import Base, Member import discord import asyncio class Baydb: engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() conn = engine.connect()
from discord.ext import commands from utils import * from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from member import Base, Member from config import Config import discord import asyncio class Baydb: engine = create_engine(Config.MAIN['dbpath']) Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() conn = engine.connect()
Move another usage of DB into ini file thing.
Move another usage of DB into ini file thing.
Python
agpl-3.0
dark-echo/Bay-Oh-Woolph,freiheit/Bay-Oh-Woolph
from discord.ext import commands from utils import * from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from member import Base, Member + from config import Config import discord import asyncio class Baydb: - engine = create_engine('sqlite:///bayohwoolph.db') + engine = create_engine(Config.MAIN['dbpath']) Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() conn = engine.connect()
Move another usage of DB into ini file thing.
## Code Before: from discord.ext import commands from utils import * from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from member import Base, Member import discord import asyncio class Baydb: engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() conn = engine.connect() ## Instruction: Move another usage of DB into ini file thing. ## Code After: from discord.ext import commands from utils import * from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from member import Base, Member from config import Config import discord import asyncio class Baydb: engine = create_engine(Config.MAIN['dbpath']) Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) session = DBSession() conn = engine.connect()
// ... existing code ... from sqlalchemy.orm import sessionmaker from member import Base, Member from config import Config import discord import asyncio // ... modified code ... class Baydb: engine = create_engine(Config.MAIN['dbpath']) Base.metadata.bind = engine DBSession = sessionmaker(bind=engine) // ... rest of the code ...
82dcd51c59eecccac4e7d9ee1dac754b27ff9ed2
mzalendo/feedback/views.py
mzalendo/feedback/views.py
from django.shortcuts import render_to_response from django.template import RequestContext from django.views.decorators.csrf import csrf_protect from models import Feedback from forms import FeedbackForm @csrf_protect def add(request): """Gather feedback for a page, and if it is ok show a thanks message and link back to the page.""" submit_was_success = False return_to_url = None # If it is a post request try to create the feedback if request.method == 'POST': form = FeedbackForm( request.POST ) if form.is_valid(): feedback = Feedback() feedback.url = form.cleaned_data['url'] feedback.email = form.cleaned_data['email'] feedback.comment = form.cleaned_data['comment'] # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: feedback.status = 'spammy' if request.user.is_authenticated(): feedback.user = request.user feedback.save() submit_was_success = True return_to_url = feedback.url or None else: # use GET to grab the url if set form = FeedbackForm(initial=request.GET) return render_to_response( 'feedback/add.html', { 'form': form, 'submit_was_success': submit_was_success, 'return_to_url': return_to_url, }, context_instance=RequestContext(request) )
from django.shortcuts import render_to_response from django.template import RequestContext from django.views.decorators.csrf import csrf_protect from models import Feedback from forms import FeedbackForm import re @csrf_protect def add(request): """Gather feedback for a page, and if it is ok show a thanks message and link back to the page.""" submit_was_success = False return_to_url = None # If it is a post request try to create the feedback if request.method == 'POST': form = FeedbackForm( request.POST ) if form.is_valid(): feedback = Feedback() feedback.url = form.cleaned_data['url'] feedback.email = form.cleaned_data['email'] feedback.comment = form.cleaned_data['comment'] # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: feedback.status = 'spammy' # if the comment starts with an html tag it is probably spam if re.search('\A\s*<\w+>', form.cleaned_data['comment']): feedback.status = 'spammy' if request.user.is_authenticated(): feedback.user = request.user feedback.save() submit_was_success = True return_to_url = feedback.url or None else: # use GET to grab the url if set form = FeedbackForm(initial=request.GET) return render_to_response( 'feedback/add.html', { 'form': form, 'submit_was_success': submit_was_success, 'return_to_url': return_to_url, }, context_instance=RequestContext(request) )
Mark feedback comments starting with a html tag as spammy
Mark feedback comments starting with a html tag as spammy
Python
agpl-3.0
ken-muturi/pombola,mysociety/pombola,hzj123/56th,hzj123/56th,mysociety/pombola,ken-muturi/pombola,hzj123/56th,Hutspace/odekro,geoffkilpin/pombola,patricmutwiri/pombola,mysociety/pombola,patricmutwiri/pombola,patricmutwiri/pombola,hzj123/56th,Hutspace/odekro,ken-muturi/pombola,Hutspace/odekro,patricmutwiri/pombola,Hutspace/odekro,mysociety/pombola,ken-muturi/pombola,geoffkilpin/pombola,geoffkilpin/pombola,patricmutwiri/pombola,hzj123/56th,geoffkilpin/pombola,ken-muturi/pombola,mysociety/pombola,patricmutwiri/pombola,geoffkilpin/pombola,hzj123/56th,Hutspace/odekro,geoffkilpin/pombola,mysociety/pombola,ken-muturi/pombola
from django.shortcuts import render_to_response from django.template import RequestContext from django.views.decorators.csrf import csrf_protect from models import Feedback from forms import FeedbackForm + + import re + @csrf_protect def add(request): """Gather feedback for a page, and if it is ok show a thanks message and link back to the page.""" submit_was_success = False return_to_url = None # If it is a post request try to create the feedback if request.method == 'POST': form = FeedbackForm( request.POST ) if form.is_valid(): feedback = Feedback() feedback.url = form.cleaned_data['url'] feedback.email = form.cleaned_data['email'] feedback.comment = form.cleaned_data['comment'] # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: + feedback.status = 'spammy' + + # if the comment starts with an html tag it is probably spam + if re.search('\A\s*<\w+>', form.cleaned_data['comment']): feedback.status = 'spammy' if request.user.is_authenticated(): feedback.user = request.user feedback.save() submit_was_success = True return_to_url = feedback.url or None else: # use GET to grab the url if set form = FeedbackForm(initial=request.GET) return render_to_response( 'feedback/add.html', { 'form': form, 'submit_was_success': submit_was_success, 'return_to_url': return_to_url, }, context_instance=RequestContext(request) )
Mark feedback comments starting with a html tag as spammy
## Code Before: from django.shortcuts import render_to_response from django.template import RequestContext from django.views.decorators.csrf import csrf_protect from models import Feedback from forms import FeedbackForm @csrf_protect def add(request): """Gather feedback for a page, and if it is ok show a thanks message and link back to the page.""" submit_was_success = False return_to_url = None # If it is a post request try to create the feedback if request.method == 'POST': form = FeedbackForm( request.POST ) if form.is_valid(): feedback = Feedback() feedback.url = form.cleaned_data['url'] feedback.email = form.cleaned_data['email'] feedback.comment = form.cleaned_data['comment'] # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: feedback.status = 'spammy' if request.user.is_authenticated(): feedback.user = request.user feedback.save() submit_was_success = True return_to_url = feedback.url or None else: # use GET to grab the url if set form = FeedbackForm(initial=request.GET) return render_to_response( 'feedback/add.html', { 'form': form, 'submit_was_success': submit_was_success, 'return_to_url': return_to_url, }, context_instance=RequestContext(request) ) ## Instruction: Mark feedback comments starting with a html tag as spammy ## Code After: from django.shortcuts import render_to_response from django.template import RequestContext from django.views.decorators.csrf import csrf_protect from models import Feedback from forms import FeedbackForm import re @csrf_protect def add(request): """Gather feedback for a page, and if it is ok show a thanks message and link back to the page.""" submit_was_success = False return_to_url = None # If it is a post request try to create the feedback if request.method == 'POST': form = FeedbackForm( request.POST ) if form.is_valid(): feedback = Feedback() feedback.url = form.cleaned_data['url'] feedback.email = form.cleaned_data['email'] feedback.comment = form.cleaned_data['comment'] # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: feedback.status = 'spammy' # if the comment starts with an html tag it is probably spam if re.search('\A\s*<\w+>', form.cleaned_data['comment']): feedback.status = 'spammy' if request.user.is_authenticated(): feedback.user = request.user feedback.save() submit_was_success = True return_to_url = feedback.url or None else: # use GET to grab the url if set form = FeedbackForm(initial=request.GET) return render_to_response( 'feedback/add.html', { 'form': form, 'submit_was_success': submit_was_success, 'return_to_url': return_to_url, }, context_instance=RequestContext(request) )
// ... existing code ... from models import Feedback from forms import FeedbackForm import re @csrf_protect // ... modified code ... # if there is any content in the honeypot field then label this comment as spammy if form.cleaned_data['website']: feedback.status = 'spammy' # if the comment starts with an html tag it is probably spam if re.search('\A\s*<\w+>', form.cleaned_data['comment']): feedback.status = 'spammy' // ... rest of the code ...
c98ab8807440e3cdbb98e11c53c7f246c35614fe
dedupe/convenience.py
dedupe/convenience.py
import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' random_pairs = dedupe.core.randomPairs(len(data), sample_size) return tuple((data[k1], data[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' data_list = data.values() random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
Change dataSample to generate indices of random pair using list of values
Change dataSample to generate indices of random pair using list of values
Python
mit
nmiranda/dedupe,01-/dedupe,neozhangthe1/dedupe,neozhangthe1/dedupe,nmiranda/dedupe,davidkunio/dedupe,dedupeio/dedupe,dedupeio/dedupe-examples,datamade/dedupe,tfmorris/dedupe,tfmorris/dedupe,davidkunio/dedupe,01-/dedupe,datamade/dedupe,pombredanne/dedupe,dedupeio/dedupe,pombredanne/dedupe
import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' + data_list = data.values() + random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) - random_pairs = dedupe.core.randomPairs(len(data), sample_size) - - return tuple((data[k1], data[k2]) for k1, k2 in random_pairs) + return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
Change dataSample to generate indices of random pair using list of values
## Code Before: import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' random_pairs = dedupe.core.randomPairs(len(data), sample_size) return tuple((data[k1], data[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records ## Instruction: Change dataSample to generate indices of random pair using list of values ## Code After: import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' data_list = data.values() random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
// ... existing code ... '''Randomly sample pairs of records from a data dictionary''' data_list = data.values() random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) // ... rest of the code ...
4aa6714284cb45a2747cea8e0f38e8fbcd8ec0bc
pymatgen/core/design_patterns.py
pymatgen/core/design_patterns.py
from __future__ import division, unicode_literals """ This module defines some useful design patterns. """ __author__ = "Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "1.0" __maintainer__ = "Shyue Ping Ong" __email__ = "[email protected]" __status__ = "Production" __date__ = "Sep 23, 2011" class Enum(set): """ Creates an enum out of a set. """ def __getattr__(self, name): if name in self: return name raise AttributeError class NullFile(object): """A file object that is associated to /dev/null.""" def __new__(cls): import os return open(os.devnull, 'w') def __init__(self): """no-op""" class NullStream(object): """A fake stream with a no-op write..""" def write(*args): """no-op"""
from __future__ import division, unicode_literals """ This module defines some useful design patterns. """ __author__ = "Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "1.0" __maintainer__ = "Shyue Ping Ong" __email__ = "[email protected]" __status__ = "Production" __date__ = "Sep 23, 2011" class Enum(set): """ Creates an enum out of a set. """ def __getattr__(self, name): if name in self: return name raise AttributeError
Move NullFile and NullStream to monty
Move NullFile and NullStream to monty
Python
mit
Bismarrck/pymatgen,Bismarrck/pymatgen,sonium0/pymatgen,rousseab/pymatgen,Dioptas/pymatgen,migueldiascosta/pymatgen,yanikou19/pymatgen,ctoher/pymatgen,migueldiascosta/pymatgen,yanikou19/pymatgen,rousseab/pymatgen,sonium0/pymatgen,ctoher/pymatgen,ctoher/pymatgen,rousseab/pymatgen,sonium0/pymatgen,Bismarrck/pymatgen,migueldiascosta/pymatgen,Bismarrck/pymatgen,Bismarrck/pymatgen,yanikou19/pymatgen,Dioptas/pymatgen
from __future__ import division, unicode_literals """ This module defines some useful design patterns. """ __author__ = "Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "1.0" __maintainer__ = "Shyue Ping Ong" __email__ = "[email protected]" __status__ = "Production" __date__ = "Sep 23, 2011" class Enum(set): """ Creates an enum out of a set. """ def __getattr__(self, name): if name in self: return name raise AttributeError - class NullFile(object): - """A file object that is associated to /dev/null.""" - def __new__(cls): - import os - return open(os.devnull, 'w') - - def __init__(self): - """no-op""" - - - class NullStream(object): - """A fake stream with a no-op write..""" - def write(*args): - """no-op""" - -
Move NullFile and NullStream to monty
## Code Before: from __future__ import division, unicode_literals """ This module defines some useful design patterns. """ __author__ = "Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "1.0" __maintainer__ = "Shyue Ping Ong" __email__ = "[email protected]" __status__ = "Production" __date__ = "Sep 23, 2011" class Enum(set): """ Creates an enum out of a set. """ def __getattr__(self, name): if name in self: return name raise AttributeError class NullFile(object): """A file object that is associated to /dev/null.""" def __new__(cls): import os return open(os.devnull, 'w') def __init__(self): """no-op""" class NullStream(object): """A fake stream with a no-op write..""" def write(*args): """no-op""" ## Instruction: Move NullFile and NullStream to monty ## Code After: from __future__ import division, unicode_literals """ This module defines some useful design patterns. """ __author__ = "Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "1.0" __maintainer__ = "Shyue Ping Ong" __email__ = "[email protected]" __status__ = "Production" __date__ = "Sep 23, 2011" class Enum(set): """ Creates an enum out of a set. """ def __getattr__(self, name): if name in self: return name raise AttributeError
// ... existing code ... raise AttributeError // ... rest of the code ...
5178318df905ed1a68d312adb3936e8748789b2b
tests/test_views.py
tests/test_views.py
import json import unittest from mock import patch from watchman import views class TestWatchman(unittest.TestCase): def setUp(self): pass @patch('watchman.views.check_databases') def test_response_content_type_json(self, patched_check_databases): patched_check_databases.return_value = [] response = views.status('') self.assertEqual(response['Content-Type'], 'application/json') @patch('watchman.views.check_databases') def test_response_contains_expected_checks(self, patched_check_databases): expected_checks = ['databases'] patched_check_databases.return_value = [] response = views.status('') content = json.loads(response.content) self.assertItemsEqual(expected_checks, content.keys()) def tearDown(self): pass
import json import unittest from mock import patch from watchman import views class TestWatchman(unittest.TestCase): def setUp(self): pass @patch('watchman.views.check_databases') def test_response_content_type_json(self, patched_check_databases): patched_check_databases.return_value = [] response = views.status('') self.assertEqual(response['Content-Type'], 'application/json') @patch('watchman.views.check_databases') def test_response_contains_expected_checks(self, patched_check_databases): expected_checks = ['databases'] patched_check_databases.return_value = [] response = views.status('') content = json.loads(response.content) self.assertItemsEqual(expected_checks, content.keys()) def test_check_database_handles_exception(self): response = views.check_database('foo') self.assertFalse(response['foo']['ok']) self.assertEqual(response['foo']['error'], "The connection foo doesn't exist") def tearDown(self): pass
Test exception handling in `check_database`
Test exception handling in `check_database`
Python
bsd-3-clause
JBKahn/django-watchman,mwarkentin/django-watchman,mwarkentin/django-watchman,ulope/django-watchman,gerlachry/django-watchman,blag/django-watchman,JBKahn/django-watchman,blag/django-watchman,gerlachry/django-watchman,ulope/django-watchman
import json import unittest from mock import patch from watchman import views class TestWatchman(unittest.TestCase): def setUp(self): pass @patch('watchman.views.check_databases') def test_response_content_type_json(self, patched_check_databases): patched_check_databases.return_value = [] response = views.status('') self.assertEqual(response['Content-Type'], 'application/json') @patch('watchman.views.check_databases') def test_response_contains_expected_checks(self, patched_check_databases): expected_checks = ['databases'] patched_check_databases.return_value = [] response = views.status('') content = json.loads(response.content) self.assertItemsEqual(expected_checks, content.keys()) + def test_check_database_handles_exception(self): + response = views.check_database('foo') + self.assertFalse(response['foo']['ok']) + self.assertEqual(response['foo']['error'], "The connection foo doesn't exist") + def tearDown(self): pass
Test exception handling in `check_database`
## Code Before: import json import unittest from mock import patch from watchman import views class TestWatchman(unittest.TestCase): def setUp(self): pass @patch('watchman.views.check_databases') def test_response_content_type_json(self, patched_check_databases): patched_check_databases.return_value = [] response = views.status('') self.assertEqual(response['Content-Type'], 'application/json') @patch('watchman.views.check_databases') def test_response_contains_expected_checks(self, patched_check_databases): expected_checks = ['databases'] patched_check_databases.return_value = [] response = views.status('') content = json.loads(response.content) self.assertItemsEqual(expected_checks, content.keys()) def tearDown(self): pass ## Instruction: Test exception handling in `check_database` ## Code After: import json import unittest from mock import patch from watchman import views class TestWatchman(unittest.TestCase): def setUp(self): pass @patch('watchman.views.check_databases') def test_response_content_type_json(self, patched_check_databases): patched_check_databases.return_value = [] response = views.status('') self.assertEqual(response['Content-Type'], 'application/json') @patch('watchman.views.check_databases') def test_response_contains_expected_checks(self, patched_check_databases): expected_checks = ['databases'] patched_check_databases.return_value = [] response = views.status('') content = json.loads(response.content) self.assertItemsEqual(expected_checks, content.keys()) def test_check_database_handles_exception(self): response = views.check_database('foo') self.assertFalse(response['foo']['ok']) self.assertEqual(response['foo']['error'], "The connection foo doesn't exist") def tearDown(self): pass
// ... existing code ... self.assertItemsEqual(expected_checks, content.keys()) def test_check_database_handles_exception(self): response = views.check_database('foo') self.assertFalse(response['foo']['ok']) self.assertEqual(response['foo']['error'], "The connection foo doesn't exist") def tearDown(self): pass // ... rest of the code ...
80e4caad24bceabd8e15133a96a6aaddd9a97c07
code/type_null_true_false.py
code/type_null_true_false.py
def if_value(values): print('"if value":') for k, v in values: print("%s - %s" % (k, 'true' if v else 'false')) print() def nil_value(values): print('"if value is None":') for k, v in values: print("%s - %s" % (k, 'true' if v is None else 'false')) print() def empty_value(values): print('"if len(value)":') for k, v in values: try: print("%s - %s" % (k, 'true' if len(v) else 'false')) except TypeError as e: print("%s - %s" % (k, e)) values = [ ("'string'", 'string'), ("''", ''), ('[1, 2, 3]', [1, 2, 3]), ('[]', []), ('5', 5), ('0', 0), (True, True), (False, False), (None, None), ] if_value(values) nil_value(values) empty_value(values)
def check(label, fn, values): print(label) for value in values: try: result = 'true' if fn(value) else 'false' except TypeError as e: result = 'error: %s' % e print(" %-9r - %s" % (value, result)) print() values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None] check('if value:', lambda v: v, values) check('if value is None:', lambda v: v is None, values) check('if len(value):', lambda v: len(v), values)
Refactor Null/True/False to look more pythonic
Refactor Null/True/False to look more pythonic
Python
mit
evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,evmorov/lang-compare
- def if_value(values): - print('"if value":') + def check(label, fn, values): + print(label) - for k, v in values: + for value in values: - print("%s - %s" % (k, 'true' if v else 'false')) + try: + result = 'true' if fn(value) else 'false' + except TypeError as e: + result = 'error: %s' % e + print(" %-9r - %s" % (value, result)) - print() + print() + values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None] - def nil_value(values): - print('"if value is None":') - for k, v in values: - print("%s - %s" % (k, 'true' if v is None else 'false')) - print() + check('if value:', lambda v: v, values) + check('if value is None:', lambda v: v is None, values) + check('if len(value):', lambda v: len(v), values) - def empty_value(values): - print('"if len(value)":') - for k, v in values: - try: - print("%s - %s" % (k, 'true' if len(v) else 'false')) - except TypeError as e: - print("%s - %s" % (k, e)) - values = [ - ("'string'", 'string'), - ("''", ''), - ('[1, 2, 3]', [1, 2, 3]), - ('[]', []), - ('5', 5), - ('0', 0), - (True, True), - (False, False), - (None, None), - ] - - if_value(values) - nil_value(values) - empty_value(values) -
Refactor Null/True/False to look more pythonic
## Code Before: def if_value(values): print('"if value":') for k, v in values: print("%s - %s" % (k, 'true' if v else 'false')) print() def nil_value(values): print('"if value is None":') for k, v in values: print("%s - %s" % (k, 'true' if v is None else 'false')) print() def empty_value(values): print('"if len(value)":') for k, v in values: try: print("%s - %s" % (k, 'true' if len(v) else 'false')) except TypeError as e: print("%s - %s" % (k, e)) values = [ ("'string'", 'string'), ("''", ''), ('[1, 2, 3]', [1, 2, 3]), ('[]', []), ('5', 5), ('0', 0), (True, True), (False, False), (None, None), ] if_value(values) nil_value(values) empty_value(values) ## Instruction: Refactor Null/True/False to look more pythonic ## Code After: def check(label, fn, values): print(label) for value in values: try: result = 'true' if fn(value) else 'false' except TypeError as e: result = 'error: %s' % e print(" %-9r - %s" % (value, result)) print() values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None] check('if value:', lambda v: v, values) check('if value is None:', lambda v: v is None, values) check('if len(value):', lambda v: len(v), values)
... def check(label, fn, values): print(label) for value in values: try: result = 'true' if fn(value) else 'false' except TypeError as e: result = 'error: %s' % e print(" %-9r - %s" % (value, result)) print() values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None] check('if value:', lambda v: v, values) check('if value is None:', lambda v: v is None, values) check('if len(value):', lambda v: len(v), values) ...
04a7de877c50bc84428e7bb7d30b1c6cac00a59f
ipywidgets/widgets/tests/test_widget_selection.py
ipywidgets/widgets/tests/test_widget_selection.py
import warnings from unittest import TestCase from ipywidgets import Dropdown class TestDropdown(TestCase): def test_construction(self): Dropdown() def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: warnings.simplefilter("module") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) assert len(w) == 1 assert issubclass(w[-1].category, DeprecationWarning) assert "deprecated" in str(w[-1].message)
import warnings from unittest import TestCase from ipywidgets import Dropdown class TestDropdown(TestCase): def test_construction(self): Dropdown() def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: warnings.simplefilter("always") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) assert len(w) > 0 assert issubclass(w[-1].category, DeprecationWarning) assert "Support for mapping types has been deprecated" in str(w[-1].message)
Use simplefilter('always') for testing the warning
Use simplefilter('always') for testing the warning * Use `warnings.simplefilter('always')` for DeprecationWarning * More specific test on warning message
Python
bsd-3-clause
jupyter-widgets/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets,jupyter-widgets/ipywidgets,ipython/ipywidgets,jupyter-widgets/ipywidgets,ipython/ipywidgets,jupyter-widgets/ipywidgets,SylvainCorlay/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets,SylvainCorlay/ipywidgets,ipython/ipywidgets
import warnings from unittest import TestCase from ipywidgets import Dropdown class TestDropdown(TestCase): def test_construction(self): Dropdown() def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: - warnings.simplefilter("module") + warnings.simplefilter("always") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) - assert len(w) == 1 + assert len(w) > 0 assert issubclass(w[-1].category, DeprecationWarning) - assert "deprecated" in str(w[-1].message) + assert "Support for mapping types has been deprecated" in str(w[-1].message)
Use simplefilter('always') for testing the warning
## Code Before: import warnings from unittest import TestCase from ipywidgets import Dropdown class TestDropdown(TestCase): def test_construction(self): Dropdown() def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: warnings.simplefilter("module") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) assert len(w) == 1 assert issubclass(w[-1].category, DeprecationWarning) assert "deprecated" in str(w[-1].message) ## Instruction: Use simplefilter('always') for testing the warning ## Code After: import warnings from unittest import TestCase from ipywidgets import Dropdown class TestDropdown(TestCase): def test_construction(self): Dropdown() def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: warnings.simplefilter("always") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) assert len(w) > 0 assert issubclass(w[-1].category, DeprecationWarning) assert "Support for mapping types has been deprecated" in str(w[-1].message)
// ... existing code ... def test_deprecation_warning_mapping_options(self): with warnings.catch_warnings(record=True) as w: warnings.simplefilter("always") Dropdown(options={'One': 1, 'Two': 2, 'Three': 3}) assert len(w) > 0 assert issubclass(w[-1].category, DeprecationWarning) assert "Support for mapping types has been deprecated" in str(w[-1].message) // ... rest of the code ...
4425aa1170a1acd3ed69c32ba5e3885301593524
salt/returners/redis_return.py
salt/returners/redis_return.py
''' Return data to a redis server To enable this returner the minion will need the python client for redis installed and the following values configured in the minion or master config, these are the defaults: redis.db: '0' redis.host: 'salt' redis.port: 6379 ''' # Import python libs import json try: import redis has_redis = True except ImportError: has_redis = False def __virtual__(): if not has_redis: return False return 'redis_return' def _get_serv(): ''' Return a redis server object ''' return redis.Redis( host=__salt__['config.option']('redis.host'), port=__salt__['config.option']('redis.port'), db=__salt__['config.option']('redis.db')) def returner(ret): ''' Return data to a redis data store ''' serv = _get_serv() serv.sadd('{0}:jobs'.format(ret['id'])) serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return']))) serv.sadd('jobs', ret['jid']) serv.sadd(ret['jid'], ret['id'])
''' Return data to a redis server To enable this returner the minion will need the python client for redis installed and the following values configured in the minion or master config, these are the defaults: redis.db: '0' redis.host: 'salt' redis.port: 6379 ''' # Import python libs import json try: import redis has_redis = True except ImportError: has_redis = False def __virtual__(): if not has_redis: return False return 'redis_return' def _get_serv(): ''' Return a redis server object ''' return redis.Redis( host=__salt__['config.option']('redis.host'), port=__salt__['config.option']('redis.port'), db=__salt__['config.option']('redis.db')) def returner(ret): ''' Return data to a redis data store ''' serv = _get_serv() serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret)) serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid']) serv.sadd('minions', ret['id'])
Restructure redis returner, since it did notwork before anyway
Restructure redis returner, since it did notwork before anyway
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
''' Return data to a redis server To enable this returner the minion will need the python client for redis installed and the following values configured in the minion or master config, these are the defaults: redis.db: '0' redis.host: 'salt' redis.port: 6379 ''' # Import python libs import json try: import redis has_redis = True except ImportError: has_redis = False def __virtual__(): if not has_redis: return False return 'redis_return' def _get_serv(): ''' Return a redis server object ''' return redis.Redis( host=__salt__['config.option']('redis.host'), port=__salt__['config.option']('redis.port'), db=__salt__['config.option']('redis.db')) def returner(ret): ''' Return data to a redis data store ''' serv = _get_serv() - serv.sadd('{0}:jobs'.format(ret['id'])) - serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return']))) + serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret)) + serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid']) - serv.sadd('jobs', ret['jid']) + serv.sadd('minions', ret['id']) - serv.sadd(ret['jid'], ret['id'])
Restructure redis returner, since it did notwork before anyway
## Code Before: ''' Return data to a redis server To enable this returner the minion will need the python client for redis installed and the following values configured in the minion or master config, these are the defaults: redis.db: '0' redis.host: 'salt' redis.port: 6379 ''' # Import python libs import json try: import redis has_redis = True except ImportError: has_redis = False def __virtual__(): if not has_redis: return False return 'redis_return' def _get_serv(): ''' Return a redis server object ''' return redis.Redis( host=__salt__['config.option']('redis.host'), port=__salt__['config.option']('redis.port'), db=__salt__['config.option']('redis.db')) def returner(ret): ''' Return data to a redis data store ''' serv = _get_serv() serv.sadd('{0}:jobs'.format(ret['id'])) serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return']))) serv.sadd('jobs', ret['jid']) serv.sadd(ret['jid'], ret['id']) ## Instruction: Restructure redis returner, since it did notwork before anyway ## Code After: ''' Return data to a redis server To enable this returner the minion will need the python client for redis installed and the following values configured in the minion or master config, these are the defaults: redis.db: '0' redis.host: 'salt' redis.port: 6379 ''' # Import python libs import json try: import redis has_redis = True except ImportError: has_redis = False def __virtual__(): if not has_redis: return False return 'redis_return' def _get_serv(): ''' Return a redis server object ''' return redis.Redis( host=__salt__['config.option']('redis.host'), port=__salt__['config.option']('redis.port'), db=__salt__['config.option']('redis.db')) def returner(ret): ''' Return data to a redis data store ''' serv = _get_serv() serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret)) serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid']) serv.sadd('minions', ret['id'])
// ... existing code ... ''' serv = _get_serv() serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret)) serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid']) serv.sadd('minions', ret['id']) // ... rest of the code ...
6a4046aafe43930c202e2f18a55b1cd8517d95f9
testanalyzer/javaanalyzer.py
testanalyzer/javaanalyzer.py
import re from fileanalyzer import FileAnalyzer class JavaAnalyzer(FileAnalyzer): def get_class_count(self, content): return len( re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content)) # TODO: Accept angle brackets and decline "else if" def get_function_count(self, content): return len( re.findall( "[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{", content))
import re from fileanalyzer import FileAnalyzer class JavaAnalyzer(FileAnalyzer): def get_class_count(self, content): return len( re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content)) def get_function_count(self, content): matches = re.findall( "[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{", content) matches = [ m for m in matches if "if " not in m.strip() and "if(" not in m.strip() ] return len(matches)
Fix regex to match generics
Fix regex to match generics
Python
mpl-2.0
CheriPai/TestAnalyzer,CheriPai/TestAnalyzer,CheriPai/TestAnalyzer
import re from fileanalyzer import FileAnalyzer class JavaAnalyzer(FileAnalyzer): def get_class_count(self, content): return len( - re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content)) + re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content)) - # TODO: Accept angle brackets and decline "else if" def get_function_count(self, content): - return len( - re.findall( + matches = re.findall( - "[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{", + "[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{", - content)) + content) + matches = [ + m for m in matches + if "if " not in m.strip() and "if(" not in m.strip() + ] + return len(matches)
Fix regex to match generics
## Code Before: import re from fileanalyzer import FileAnalyzer class JavaAnalyzer(FileAnalyzer): def get_class_count(self, content): return len( re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content)) # TODO: Accept angle brackets and decline "else if" def get_function_count(self, content): return len( re.findall( "[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{", content)) ## Instruction: Fix regex to match generics ## Code After: import re from fileanalyzer import FileAnalyzer class JavaAnalyzer(FileAnalyzer): def get_class_count(self, content): return len( re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content)) def get_function_count(self, content): matches = re.findall( "[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{", content) matches = [ m for m in matches if "if " not in m.strip() and "if(" not in m.strip() ] return len(matches)
# ... existing code ... def get_class_count(self, content): return len( re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content)) def get_function_count(self, content): matches = re.findall( "[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{", content) matches = [ m for m in matches if "if " not in m.strip() and "if(" not in m.strip() ] return len(matches) # ... rest of the code ...
133617660fe96a817b47d4d0fba4cfa7567dcafb
exceptional.py
exceptional.py
"""A module to demonstrate exceptions.""" import sys def convert(item): ''' Convert to an integer. Args: item: some object Returns: an integer representation of the object Throws: a ValueException ''' try: x = int(item) print(str.format('Conversion succeeded! x= {}', x)) except ValueError: print('Conversion Failed') x = -1 return x if __name__ == '__main__': print(convert(sys.argv[1]))
"""A module to demonstrate exceptions.""" import sys def convert(item): """ Convert to an integer. Args: item: some object Returns: an integer representation of the object Throws: a ValueException """ try: return int(item) except (ValueError, TypeError): return -1 if __name__ == '__main__': print(convert(sys.argv[1]))
Use two return statements and remove printing
Use two return statements and remove printing
Python
mit
kentoj/python-fundamentals
"""A module to demonstrate exceptions.""" import sys def convert(item): - ''' + """ Convert to an integer. Args: item: some object Returns: an integer representation of the object Throws: a ValueException - ''' + """ try: - x = int(item) + return int(item) - print(str.format('Conversion succeeded! x= {}', x)) - except ValueError: + except (ValueError, TypeError): - print('Conversion Failed') - x = -1 - return x + return -1 if __name__ == '__main__': print(convert(sys.argv[1]))
Use two return statements and remove printing
## Code Before: """A module to demonstrate exceptions.""" import sys def convert(item): ''' Convert to an integer. Args: item: some object Returns: an integer representation of the object Throws: a ValueException ''' try: x = int(item) print(str.format('Conversion succeeded! x= {}', x)) except ValueError: print('Conversion Failed') x = -1 return x if __name__ == '__main__': print(convert(sys.argv[1])) ## Instruction: Use two return statements and remove printing ## Code After: """A module to demonstrate exceptions.""" import sys def convert(item): """ Convert to an integer. Args: item: some object Returns: an integer representation of the object Throws: a ValueException """ try: return int(item) except (ValueError, TypeError): return -1 if __name__ == '__main__': print(convert(sys.argv[1]))
... def convert(item): """ Convert to an integer. Args: ... Throws: a ValueException """ try: return int(item) except (ValueError, TypeError): return -1 if __name__ == '__main__': ...
8608283592338960c80113ff4d68f42936ddb969
linter.py
linter.py
"""This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd
"""This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
Clean up include dir code
Clean up include dir code
Python
mit
oschwald/SublimeLinter-perl
"""This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' - base_cmd = ('perl -c') + regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ - full_cmd = self.base_cmd + command = [self.executable_path, '-c'] - settings = self.get_view_settings() + include_dirs = self.get_view_settings().get('include_dirs', []) - include_dirs = settings.get('include_dirs', []) + for e in include_dirs: + command.append('-I') + command.append(shlex.quote(e)) + return command - if include_dirs: - full_cmd += ' ' . join([' -I ' + shlex.quote(include) - for include in include_dirs]) - return full_cmd -
Clean up include dir code
## Code Before: """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd ## Instruction: Clean up include dir code ## Code After: """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
... syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR ... """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command ...
403f23ae486c14066e0a93c7deca91c5fbc15b87
plugins/brian.py
plugins/brian.py
"""Displays a randomly generated witticism from Brian Chu himself.""" import json import random __match__ = r"!brian" with open('plugins/brian_corpus/cache.json', 'r') as infile: cache = json.load(infile) with open('plugins/brian_corpus/phrases.json', 'r') as infile: phrases = json.load(infile) def generate_phrase(phrases, cache): seed_phrase = [] while len(seed_phrase) < 3: seed_phrase = random.choice(phrases).split() w1, w2 = seed_phrase[:2] chosen = [w1, w2] while "{}|{}".format(w1, w2) in cache: choice = random.choice(cache["{}|{}".format(w1, w2)]) w1, w2 = w2, choice chosen.append(choice) return ' '.join(chosen) def on_message(bot, channel, user, message): return '> {} ~brian'.format(generate_phrase(phrases, cache))
"""Displays a randomly generated witticism from Brian Chu himself.""" import json import random __match__ = r"!brian" attribution = [ "salad master", "esquire", "the one and only", "startup enthusiast", "boba king", "not-dictator", "normal citizen", "ping-pong expert" ] with open('plugins/brian_corpus/phrases.json', 'r') as infile: phrases = json.load(infile) with open('plugins/brian_corpus/cache.json', 'r') as infile: cache = json.load(infile) def generate_phrase(phrases, cache, max_length=40): seed_phrase = [] while len(seed_phrase) < 2: seed_phrase = random.choice(phrases).split() w1, = seed_phrase[:1] chosen = [w1] while w1 in cache and len(chosen)<max_length: w1 = random.choice(cache[w1]) chosen.append(w1) return ' '.join(chosen) def on_message(bot, channel, user, message): return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache), random.choice(attribution))
Use bigrams in Markov chain generator
Use bigrams in Markov chain generator
Python
mit
kvchen/keffbot,kvchen/keffbot-py
"""Displays a randomly generated witticism from Brian Chu himself.""" import json import random __match__ = r"!brian" + attribution = [ + "salad master", + "esquire", + "the one and only", + "startup enthusiast", + "boba king", + "not-dictator", + "normal citizen", + "ping-pong expert" + ] + + + with open('plugins/brian_corpus/phrases.json', 'r') as infile: + phrases = json.load(infile) with open('plugins/brian_corpus/cache.json', 'r') as infile: cache = json.load(infile) - with open('plugins/brian_corpus/phrases.json', 'r') as infile: - phrases = json.load(infile) - - def generate_phrase(phrases, cache): + def generate_phrase(phrases, cache, max_length=40): seed_phrase = [] - while len(seed_phrase) < 3: + while len(seed_phrase) < 2: seed_phrase = random.choice(phrases).split() - w1, w2 = seed_phrase[:2] + w1, = seed_phrase[:1] - chosen = [w1, w2] + chosen = [w1] + while w1 in cache and len(chosen)<max_length: + w1 = random.choice(cache[w1]) - while "{}|{}".format(w1, w2) in cache: - choice = random.choice(cache["{}|{}".format(w1, w2)]) - w1, w2 = w2, choice - chosen.append(choice) + chosen.append(w1) return ' '.join(chosen) def on_message(bot, channel, user, message): - return '> {} ~brian'.format(generate_phrase(phrases, cache)) + return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache), + random.choice(attribution))
Use bigrams in Markov chain generator
## Code Before: """Displays a randomly generated witticism from Brian Chu himself.""" import json import random __match__ = r"!brian" with open('plugins/brian_corpus/cache.json', 'r') as infile: cache = json.load(infile) with open('plugins/brian_corpus/phrases.json', 'r') as infile: phrases = json.load(infile) def generate_phrase(phrases, cache): seed_phrase = [] while len(seed_phrase) < 3: seed_phrase = random.choice(phrases).split() w1, w2 = seed_phrase[:2] chosen = [w1, w2] while "{}|{}".format(w1, w2) in cache: choice = random.choice(cache["{}|{}".format(w1, w2)]) w1, w2 = w2, choice chosen.append(choice) return ' '.join(chosen) def on_message(bot, channel, user, message): return '> {} ~brian'.format(generate_phrase(phrases, cache)) ## Instruction: Use bigrams in Markov chain generator ## Code After: """Displays a randomly generated witticism from Brian Chu himself.""" import json import random __match__ = r"!brian" attribution = [ "salad master", "esquire", "the one and only", "startup enthusiast", "boba king", "not-dictator", "normal citizen", "ping-pong expert" ] with open('plugins/brian_corpus/phrases.json', 'r') as infile: phrases = json.load(infile) with open('plugins/brian_corpus/cache.json', 'r') as infile: cache = json.load(infile) def generate_phrase(phrases, cache, max_length=40): seed_phrase = [] while len(seed_phrase) < 2: seed_phrase = random.choice(phrases).split() w1, = seed_phrase[:1] chosen = [w1] while w1 in cache and len(chosen)<max_length: w1 = random.choice(cache[w1]) chosen.append(w1) return ' '.join(chosen) def on_message(bot, channel, user, message): return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache), random.choice(attribution))
... __match__ = r"!brian" attribution = [ "salad master", "esquire", "the one and only", "startup enthusiast", "boba king", "not-dictator", "normal citizen", "ping-pong expert" ] with open('plugins/brian_corpus/phrases.json', 'r') as infile: phrases = json.load(infile) with open('plugins/brian_corpus/cache.json', 'r') as infile: ... cache = json.load(infile) def generate_phrase(phrases, cache, max_length=40): seed_phrase = [] while len(seed_phrase) < 2: seed_phrase = random.choice(phrases).split() w1, = seed_phrase[:1] chosen = [w1] while w1 in cache and len(chosen)<max_length: w1 = random.choice(cache[w1]) chosen.append(w1) return ' '.join(chosen) ... def on_message(bot, channel, user, message): return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache), random.choice(attribution)) ...
16b3dc1f8c762a751e1476d679391f3bbc82cd5d
python-prefix.py
python-prefix.py
import sys import os.path import site def main(): '''\ Check if the given prefix is included in sys.path for the given python version; if not find an alternate valid prefix. Print the result to standard out. ''' if len(sys.argv) != 3: msg = 'usage: %s <prefix> <python version>\n' % \ os.path.basename(sys.argv[0]) sys.stderr.write(msg) return 1 python_prefix = sys.argv[1] python_version = sys.argv[2] path = '%s/lib/python%s' % (python_prefix, python_version) path = os.path.normpath(path) if path[-1] != '/': path = path + '/' prefix = None for p in sys.path: if p.startswith(path): prefix = path break if not prefix: prefix = site.PREFIXES[-1] sys.stdout.write('%s\n' % prefix) return 0 if __name__ == '__main__': sys.exit(main())
import sys import os.path import site def main(): '''\ Check if the given prefix is included in sys.path for the given python version; if not find an alternate valid prefix. Print the result to standard out. ''' if len(sys.argv) != 3: msg = 'usage: %s <prefix> <python version>\n' % \ os.path.basename(sys.argv[0]) sys.stderr.write(msg) return 1 python_prefix = sys.argv[1] python_version = sys.argv[2] path = '%s/lib/python%s' % (python_prefix, python_version) path = os.path.normpath(path) if path[-1] != '/': path = path + '/' prefix = None for p in sys.path: if p.startswith(path): prefix = python_prefix break if not prefix: prefix = site.PREFIXES[-1] sys.stdout.write('%s\n' % prefix) return 0 if __name__ == '__main__': sys.exit(main())
Fix typo in previous commit.
Fix typo in previous commit.
Python
bsd-2-clause
marek-sezemsky/coreemu,tectronics/coreemu,marek-sezemsky/coreemu,guidotack/coreemu,guidotack/coreemu,tectronics/coreemu,tectronics/coreemu,gregtampa/coreemu,guidotack/coreemu,gregtampa/coreemu,marek-sezemsky/coreemu,gregtampa/coreemu
import sys import os.path import site def main(): '''\ Check if the given prefix is included in sys.path for the given python version; if not find an alternate valid prefix. Print the result to standard out. ''' if len(sys.argv) != 3: msg = 'usage: %s <prefix> <python version>\n' % \ os.path.basename(sys.argv[0]) sys.stderr.write(msg) return 1 python_prefix = sys.argv[1] python_version = sys.argv[2] path = '%s/lib/python%s' % (python_prefix, python_version) path = os.path.normpath(path) if path[-1] != '/': path = path + '/' prefix = None for p in sys.path: if p.startswith(path): - prefix = path + prefix = python_prefix break if not prefix: prefix = site.PREFIXES[-1] sys.stdout.write('%s\n' % prefix) return 0 if __name__ == '__main__': sys.exit(main())
Fix typo in previous commit.
## Code Before: import sys import os.path import site def main(): '''\ Check if the given prefix is included in sys.path for the given python version; if not find an alternate valid prefix. Print the result to standard out. ''' if len(sys.argv) != 3: msg = 'usage: %s <prefix> <python version>\n' % \ os.path.basename(sys.argv[0]) sys.stderr.write(msg) return 1 python_prefix = sys.argv[1] python_version = sys.argv[2] path = '%s/lib/python%s' % (python_prefix, python_version) path = os.path.normpath(path) if path[-1] != '/': path = path + '/' prefix = None for p in sys.path: if p.startswith(path): prefix = path break if not prefix: prefix = site.PREFIXES[-1] sys.stdout.write('%s\n' % prefix) return 0 if __name__ == '__main__': sys.exit(main()) ## Instruction: Fix typo in previous commit. ## Code After: import sys import os.path import site def main(): '''\ Check if the given prefix is included in sys.path for the given python version; if not find an alternate valid prefix. Print the result to standard out. ''' if len(sys.argv) != 3: msg = 'usage: %s <prefix> <python version>\n' % \ os.path.basename(sys.argv[0]) sys.stderr.write(msg) return 1 python_prefix = sys.argv[1] python_version = sys.argv[2] path = '%s/lib/python%s' % (python_prefix, python_version) path = os.path.normpath(path) if path[-1] != '/': path = path + '/' prefix = None for p in sys.path: if p.startswith(path): prefix = python_prefix break if not prefix: prefix = site.PREFIXES[-1] sys.stdout.write('%s\n' % prefix) return 0 if __name__ == '__main__': sys.exit(main())
... for p in sys.path: if p.startswith(path): prefix = python_prefix break if not prefix: ...
524ee1cd2f56f6fe968f409d37cbd2af1621e7f3
framework/guid/model.py
framework/guid/model.py
from framework import StoredObject, fields class Guid(StoredObject): _id = fields.StringField() referent = fields.AbstractForeignField() _meta = { 'optimistic': True, } class GuidStoredObject(StoredObject): # Redirect to content using URL redirect by default redirect_mode = 'redirect' def _ensure_guid(self): """Create GUID record if current record doesn't already have one, then point GUID to self. """ # Create GUID with specified ID if ID provided if self._primary_key: # Done if GUID already exists guid = Guid.load(self._primary_key) if guid is not None: return # Create GUID guid = Guid( _id=self._primary_key, referent=self ) guid.save() # Else create GUID optimistically else: # Create GUID guid = Guid() guid.save() guid.referent = (guid._primary_key, self._name) guid.save() # Set primary key to GUID key self._primary_key = guid._primary_key def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ rv = super(GuidStoredObject, self).save(*args, **kwargs) self._ensure_guid() return rv @property def annotations(self): """ Get meta-data annotations associated with object. """ return self.metadata__annotated
from framework import StoredObject, fields class Guid(StoredObject): _id = fields.StringField() referent = fields.AbstractForeignField() _meta = { 'optimistic': True, } class GuidStoredObject(StoredObject): # Redirect to content using URL redirect by default redirect_mode = 'redirect' def _ensure_guid(self): """Create GUID record if current record doesn't already have one, then point GUID to self. """ # Create GUID with specified ID if ID provided if self._primary_key: # Done if GUID already exists guid = Guid.load(self._primary_key) if guid is not None: return # Create GUID guid = Guid( _id=self._primary_key, referent=self ) guid.save() # Else create GUID optimistically else: # Create GUID guid = Guid() guid.save() guid.referent = (guid._primary_key, self._name) guid.save() # Set primary key to GUID key self._primary_key = guid._primary_key def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ self._ensure_guid() return super(GuidStoredObject, self).save(*args, **kwargs) @property def annotations(self): """ Get meta-data annotations associated with object. """ return self.metadata__annotated
Fix last commit: Must ensure GUID before saving so that PK is defined
Fix last commit: Must ensure GUID before saving so that PK is defined
Python
apache-2.0
zkraime/osf.io,emetsger/osf.io,RomanZWang/osf.io,chennan47/osf.io,TomHeatwole/osf.io,adlius/osf.io,cwisecarver/osf.io,petermalcolm/osf.io,mfraezz/osf.io,wearpants/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,samanehsan/osf.io,CenterForOpenScience/osf.io,caseyrollins/osf.io,felliott/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,jmcarp/osf.io,jeffreyliu3230/osf.io,sbt9uc/osf.io,Ghalko/osf.io,cwisecarver/osf.io,aaxelb/osf.io,pattisdr/osf.io,SSJohns/osf.io,billyhunt/osf.io,danielneis/osf.io,GageGaskins/osf.io,ZobairAlijan/osf.io,ckc6cz/osf.io,acshi/osf.io,cslzchen/osf.io,njantrania/osf.io,samchrisinger/osf.io,acshi/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,lamdnhan/osf.io,njantrania/osf.io,billyhunt/osf.io,zamattiac/osf.io,felliott/osf.io,rdhyee/osf.io,alexschiller/osf.io,mluo613/osf.io,zachjanicki/osf.io,Nesiehr/osf.io,alexschiller/osf.io,crcresearch/osf.io,kwierman/osf.io,arpitar/osf.io,cwisecarver/osf.io,TomHeatwole/osf.io,abought/osf.io,mfraezz/osf.io,ckc6cz/osf.io,MerlinZhang/osf.io,fabianvf/osf.io,himanshuo/osf.io,mattclark/osf.io,mluo613/osf.io,asanfilippo7/osf.io,Johnetordoff/osf.io,samchrisinger/osf.io,TomBaxter/osf.io,himanshuo/osf.io,mluo613/osf.io,zkraime/osf.io,KAsante95/osf.io,revanthkolli/osf.io,himanshuo/osf.io,barbour-em/osf.io,adlius/osf.io,caneruguz/osf.io,fabianvf/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,binoculars/osf.io,doublebits/osf.io,ckc6cz/osf.io,caseyrygt/osf.io,pattisdr/osf.io,barbour-em/osf.io,laurenrevere/osf.io,revanthkolli/osf.io,jeffreyliu3230/osf.io,barbour-em/osf.io,cldershem/osf.io,monikagrabowska/osf.io,petermalcolm/osf.io,danielneis/osf.io,jinluyuan/osf.io,danielneis/osf.io,chrisseto/osf.io,lyndsysimon/osf.io,jeffreyliu3230/osf.io,jinluyuan/osf.io,jeffreyliu3230/osf.io,chennan47/osf.io,ticklemepierce/osf.io,aaxelb/osf.io,mluke93/osf.io,samchrisinger/osf.io,SSJohns/osf.io,binoculars/osf.io,reinaH/osf.io,mluo613/osf.io,cosenal/osf.io,CenterForOpenScience/osf.io,Ghalko/osf.io,caneruguz/osf.io,brandonPurvis/osf.io,GageGaskins/osf.io,DanielSBrown/osf.io,mfraezz/osf.io,bdyetton/prettychart,jinluyuan/osf.io,baylee-d/osf.io,icereval/osf.io,kushG/osf.io,reinaH/osf.io,jolene-esposito/osf.io,mattclark/osf.io,ckc6cz/osf.io,reinaH/osf.io,mattclark/osf.io,zachjanicki/osf.io,wearpants/osf.io,asanfilippo7/osf.io,cldershem/osf.io,cldershem/osf.io,laurenrevere/osf.io,lyndsysimon/osf.io,cslzchen/osf.io,zkraime/osf.io,GageGaskins/osf.io,brianjgeiger/osf.io,mluke93/osf.io,zamattiac/osf.io,GaryKriebel/osf.io,TomBaxter/osf.io,lamdnhan/osf.io,brianjgeiger/osf.io,dplorimer/osf,cosenal/osf.io,aaxelb/osf.io,erinspace/osf.io,kushG/osf.io,saradbowman/osf.io,chrisseto/osf.io,kushG/osf.io,asanfilippo7/osf.io,caseyrygt/osf.io,Ghalko/osf.io,hmoco/osf.io,acshi/osf.io,brandonPurvis/osf.io,MerlinZhang/osf.io,SSJohns/osf.io,AndrewSallans/osf.io,leb2dg/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,cslzchen/osf.io,wearpants/osf.io,DanielSBrown/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,TomBaxter/osf.io,samanehsan/osf.io,kushG/osf.io,chrisseto/osf.io,binoculars/osf.io,arpitar/osf.io,doublebits/osf.io,ZobairAlijan/osf.io,samanehsan/osf.io,dplorimer/osf,acshi/osf.io,sbt9uc/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,felliott/osf.io,amyshi188/osf.io,mluke93/osf.io,zachjanicki/osf.io,GageGaskins/osf.io,dplorimer/osf,zkraime/osf.io,jolene-esposito/osf.io,rdhyee/osf.io,amyshi188/osf.io,doublebits/osf.io,lamdnhan/osf.io,mluo613/osf.io,Nesiehr/osf.io,brandonPurvis/osf.io,arpitar/osf.io,amyshi188/osf.io,jolene-esposito/osf.io,monikagrabowska/osf.io,zachjanicki/osf.io,petermalcolm/osf.io,cslzchen/osf.io,wearpants/osf.io,cldershem/osf.io,billyhunt/osf.io,njantrania/osf.io,DanielSBrown/osf.io,adlius/osf.io,jnayak1/osf.io,samanehsan/osf.io,felliott/osf.io,doublebits/osf.io,jolene-esposito/osf.io,crcresearch/osf.io,bdyetton/prettychart,bdyetton/prettychart,barbour-em/osf.io,lamdnhan/osf.io,Ghalko/osf.io,sbt9uc/osf.io,KAsante95/osf.io,sloria/osf.io,kch8qx/osf.io,erinspace/osf.io,zamattiac/osf.io,caneruguz/osf.io,kwierman/osf.io,samchrisinger/osf.io,TomHeatwole/osf.io,emetsger/osf.io,alexschiller/osf.io,doublebits/osf.io,caseyrygt/osf.io,ticklemepierce/osf.io,Nesiehr/osf.io,jinluyuan/osf.io,cosenal/osf.io,haoyuchen1992/osf.io,jmcarp/osf.io,kch8qx/osf.io,HarryRybacki/osf.io,MerlinZhang/osf.io,icereval/osf.io,kwierman/osf.io,alexschiller/osf.io,adlius/osf.io,icereval/osf.io,leb2dg/osf.io,jmcarp/osf.io,kch8qx/osf.io,jnayak1/osf.io,hmoco/osf.io,fabianvf/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,kch8qx/osf.io,CenterForOpenScience/osf.io,HarryRybacki/osf.io,bdyetton/prettychart,dplorimer/osf,aaxelb/osf.io,njantrania/osf.io,billyhunt/osf.io,ZobairAlijan/osf.io,rdhyee/osf.io,SSJohns/osf.io,billyhunt/osf.io,chennan47/osf.io,GaryKriebel/osf.io,chrisseto/osf.io,crcresearch/osf.io,revanthkolli/osf.io,kch8qx/osf.io,jnayak1/osf.io,danielneis/osf.io,GageGaskins/osf.io,amyshi188/osf.io,RomanZWang/osf.io,abought/osf.io,lyndsysimon/osf.io,cosenal/osf.io,KAsante95/osf.io,hmoco/osf.io,brandonPurvis/osf.io,haoyuchen1992/osf.io,mluke93/osf.io,GaryKriebel/osf.io,GaryKriebel/osf.io,reinaH/osf.io,leb2dg/osf.io,caneruguz/osf.io,emetsger/osf.io,jmcarp/osf.io,leb2dg/osf.io,mfraezz/osf.io,haoyuchen1992/osf.io,brandonPurvis/osf.io,cwisecarver/osf.io,caseyrollins/osf.io,caseyrygt/osf.io,kwierman/osf.io,AndrewSallans/osf.io,haoyuchen1992/osf.io,KAsante95/osf.io,jnayak1/osf.io,arpitar/osf.io,zamattiac/osf.io,petermalcolm/osf.io,HarryRybacki/osf.io,emetsger/osf.io,RomanZWang/osf.io,sloria/osf.io,erinspace/osf.io,fabianvf/osf.io,sloria/osf.io,acshi/osf.io,caseyrollins/osf.io,DanielSBrown/osf.io,asanfilippo7/osf.io,MerlinZhang/osf.io,himanshuo/osf.io,saradbowman/osf.io,alexschiller/osf.io,brianjgeiger/osf.io,sbt9uc/osf.io,rdhyee/osf.io,abought/osf.io,HarryRybacki/osf.io,TomHeatwole/osf.io,RomanZWang/osf.io,ZobairAlijan/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,abought/osf.io,lyndsysimon/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,ticklemepierce/osf.io,revanthkolli/osf.io,KAsante95/osf.io
from framework import StoredObject, fields class Guid(StoredObject): _id = fields.StringField() referent = fields.AbstractForeignField() _meta = { 'optimistic': True, } class GuidStoredObject(StoredObject): # Redirect to content using URL redirect by default redirect_mode = 'redirect' def _ensure_guid(self): """Create GUID record if current record doesn't already have one, then point GUID to self. """ # Create GUID with specified ID if ID provided if self._primary_key: # Done if GUID already exists guid = Guid.load(self._primary_key) if guid is not None: return # Create GUID guid = Guid( _id=self._primary_key, referent=self ) guid.save() # Else create GUID optimistically else: # Create GUID guid = Guid() guid.save() guid.referent = (guid._primary_key, self._name) guid.save() # Set primary key to GUID key self._primary_key = guid._primary_key def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ - rv = super(GuidStoredObject, self).save(*args, **kwargs) self._ensure_guid() - return rv + return super(GuidStoredObject, self).save(*args, **kwargs) @property def annotations(self): """ Get meta-data annotations associated with object. """ return self.metadata__annotated
Fix last commit: Must ensure GUID before saving so that PK is defined
## Code Before: from framework import StoredObject, fields class Guid(StoredObject): _id = fields.StringField() referent = fields.AbstractForeignField() _meta = { 'optimistic': True, } class GuidStoredObject(StoredObject): # Redirect to content using URL redirect by default redirect_mode = 'redirect' def _ensure_guid(self): """Create GUID record if current record doesn't already have one, then point GUID to self. """ # Create GUID with specified ID if ID provided if self._primary_key: # Done if GUID already exists guid = Guid.load(self._primary_key) if guid is not None: return # Create GUID guid = Guid( _id=self._primary_key, referent=self ) guid.save() # Else create GUID optimistically else: # Create GUID guid = Guid() guid.save() guid.referent = (guid._primary_key, self._name) guid.save() # Set primary key to GUID key self._primary_key = guid._primary_key def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ rv = super(GuidStoredObject, self).save(*args, **kwargs) self._ensure_guid() return rv @property def annotations(self): """ Get meta-data annotations associated with object. """ return self.metadata__annotated ## Instruction: Fix last commit: Must ensure GUID before saving so that PK is defined ## Code After: from framework import StoredObject, fields class Guid(StoredObject): _id = fields.StringField() referent = fields.AbstractForeignField() _meta = { 'optimistic': True, } class GuidStoredObject(StoredObject): # Redirect to content using URL redirect by default redirect_mode = 'redirect' def _ensure_guid(self): """Create GUID record if current record doesn't already have one, then point GUID to self. """ # Create GUID with specified ID if ID provided if self._primary_key: # Done if GUID already exists guid = Guid.load(self._primary_key) if guid is not None: return # Create GUID guid = Guid( _id=self._primary_key, referent=self ) guid.save() # Else create GUID optimistically else: # Create GUID guid = Guid() guid.save() guid.referent = (guid._primary_key, self._name) guid.save() # Set primary key to GUID key self._primary_key = guid._primary_key def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ self._ensure_guid() return super(GuidStoredObject, self).save(*args, **kwargs) @property def annotations(self): """ Get meta-data annotations associated with object. """ return self.metadata__annotated
... def save(self, *args, **kwargs): """ Ensure GUID on save initialization. """ self._ensure_guid() return super(GuidStoredObject, self).save(*args, **kwargs) @property ...
73eacdde5067e60f40af000237d198748c5b3cc7
PYNWapp/PYNWsite/models.py
PYNWapp/PYNWsite/models.py
from __future__ import unicode_literals from django.db import models from django.utils import timezone # Create your models here. class Event(models.Model): name = models.CharField(max_length=200) location = models.CharField(max_length=300) event_date = models.DateTimeField('event date') description = models.TextField() def __str__(self): return self.name def is_future(self): return self.event_date > timezone.now() class Post(models.Model): title = models.CharField(max_length=100, unique=True) slug = models.SlugField(max_length=100, unique=True) body = models.TextField() posted = models.DateField(db_index=True, auto_now_add=True) category = models.ForeignKey('Category') class Category(models.Model): title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True)
from __future__ import unicode_literals from django.db import models from django.utils import timezone # Create your models here. class Event(models.Model): name = models.CharField(max_length=200) location = models.CharField(max_length=300) event_date = models.DateTimeField('event date') description = models.TextField() def __str__(self): return self.name def is_future(self): return self.event_date > timezone.now() class Post(models.Model): title = models.CharField(max_length=100, unique=True) slug = models.SlugField(max_length=100, unique=True) body = models.TextField() posted = models.DateField(db_index=True, auto_now_add=True) category = models.ForeignKey('Category') class Category(models.Model): title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True) class Meta: verbose_name_plural = 'Categories'
Fix plural name for Categories model.
Fix plural name for Categories model.
Python
mit
PythonNorthwestEngland/pynw-website,PythonNorthwestEngland/pynw-website
from __future__ import unicode_literals from django.db import models from django.utils import timezone # Create your models here. class Event(models.Model): name = models.CharField(max_length=200) location = models.CharField(max_length=300) event_date = models.DateTimeField('event date') description = models.TextField() def __str__(self): return self.name def is_future(self): return self.event_date > timezone.now() class Post(models.Model): title = models.CharField(max_length=100, unique=True) slug = models.SlugField(max_length=100, unique=True) body = models.TextField() posted = models.DateField(db_index=True, auto_now_add=True) category = models.ForeignKey('Category') class Category(models.Model): title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True) + class Meta: + verbose_name_plural = 'Categories' +
Fix plural name for Categories model.
## Code Before: from __future__ import unicode_literals from django.db import models from django.utils import timezone # Create your models here. class Event(models.Model): name = models.CharField(max_length=200) location = models.CharField(max_length=300) event_date = models.DateTimeField('event date') description = models.TextField() def __str__(self): return self.name def is_future(self): return self.event_date > timezone.now() class Post(models.Model): title = models.CharField(max_length=100, unique=True) slug = models.SlugField(max_length=100, unique=True) body = models.TextField() posted = models.DateField(db_index=True, auto_now_add=True) category = models.ForeignKey('Category') class Category(models.Model): title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True) ## Instruction: Fix plural name for Categories model. ## Code After: from __future__ import unicode_literals from django.db import models from django.utils import timezone # Create your models here. class Event(models.Model): name = models.CharField(max_length=200) location = models.CharField(max_length=300) event_date = models.DateTimeField('event date') description = models.TextField() def __str__(self): return self.name def is_future(self): return self.event_date > timezone.now() class Post(models.Model): title = models.CharField(max_length=100, unique=True) slug = models.SlugField(max_length=100, unique=True) body = models.TextField() posted = models.DateField(db_index=True, auto_now_add=True) category = models.ForeignKey('Category') class Category(models.Model): title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True) class Meta: verbose_name_plural = 'Categories'
... title = models.CharField(max_length=100, db_index=True) slug = models.SlugField(max_length=100, db_index=True) class Meta: verbose_name_plural = 'Categories' ...
109b753c807dae30ee736a6f071a058fa8b68d92
tests/scoring_engine/web/views/test_services.py
tests/scoring_engine/web/views/test_services.py
from tests.scoring_engine.web.web_test import WebTest class TestServices(WebTest): def test_auth_required_services(self): self.verify_auth_required('/services') def test_auth_required_service_id(self): self.verify_auth_required('/service/1')
from tests.scoring_engine.web.web_test import WebTest from tests.scoring_engine.helpers import generate_sample_model_tree class TestServices(WebTest): def set_team_color(self, team, color): team.color = color self.session.add(team) self.session.commit() def set_blue_team(self, team): self.set_team_color(team, 'Blue') def set_white_team(self, team): self.set_team_color(team, 'White') def test_auth_required_services(self): self.verify_auth_required('/services') def test_auth_required_service_id(self): self.verify_auth_required('/service/1') def test_normal_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 200 def test_unauthorized_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_white_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 302 def test_normal_service_id(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 200 def test_unauthorized_service_id(self): self.create_default_user() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 302
Update tests for services view
Update tests for services view
Python
mit
pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine
from tests.scoring_engine.web.web_test import WebTest + from tests.scoring_engine.helpers import generate_sample_model_tree class TestServices(WebTest): + + def set_team_color(self, team, color): + team.color = color + self.session.add(team) + self.session.commit() + + def set_blue_team(self, team): + self.set_team_color(team, 'Blue') + + def set_white_team(self, team): + self.set_team_color(team, 'White') def test_auth_required_services(self): self.verify_auth_required('/services') def test_auth_required_service_id(self): self.verify_auth_required('/service/1') + def test_normal_services(self): + user = self.create_default_user() + service = generate_sample_model_tree('Service', self.session) + self.set_blue_team(user.team) + service.team = user.team + self.session.add(service) + self.session.commit() + resp = self.auth_and_get_path('/services') + assert resp.status_code == 200 + + def test_unauthorized_services(self): + user = self.create_default_user() + service = generate_sample_model_tree('Service', self.session) + self.set_white_team(user.team) + service.team = user.team + self.session.add(service) + self.session.commit() + resp = self.auth_and_get_path('/services') + assert resp.status_code == 302 + + def test_normal_service_id(self): + user = self.create_default_user() + service = generate_sample_model_tree('Service', self.session) + self.set_blue_team(user.team) + service.team = user.team + self.session.add(service) + self.session.commit() + resp = self.auth_and_get_path('/service/1') + assert resp.status_code == 200 + + def test_unauthorized_service_id(self): + self.create_default_user() + resp = self.auth_and_get_path('/service/1') + assert resp.status_code == 302
Update tests for services view
## Code Before: from tests.scoring_engine.web.web_test import WebTest class TestServices(WebTest): def test_auth_required_services(self): self.verify_auth_required('/services') def test_auth_required_service_id(self): self.verify_auth_required('/service/1') ## Instruction: Update tests for services view ## Code After: from tests.scoring_engine.web.web_test import WebTest from tests.scoring_engine.helpers import generate_sample_model_tree class TestServices(WebTest): def set_team_color(self, team, color): team.color = color self.session.add(team) self.session.commit() def set_blue_team(self, team): self.set_team_color(team, 'Blue') def set_white_team(self, team): self.set_team_color(team, 'White') def test_auth_required_services(self): self.verify_auth_required('/services') def test_auth_required_service_id(self): self.verify_auth_required('/service/1') def test_normal_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 200 def test_unauthorized_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_white_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 302 def test_normal_service_id(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 200 def test_unauthorized_service_id(self): self.create_default_user() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 302
... from tests.scoring_engine.web.web_test import WebTest from tests.scoring_engine.helpers import generate_sample_model_tree class TestServices(WebTest): def set_team_color(self, team, color): team.color = color self.session.add(team) self.session.commit() def set_blue_team(self, team): self.set_team_color(team, 'Blue') def set_white_team(self, team): self.set_team_color(team, 'White') def test_auth_required_services(self): ... def test_auth_required_service_id(self): self.verify_auth_required('/service/1') def test_normal_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 200 def test_unauthorized_services(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_white_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/services') assert resp.status_code == 302 def test_normal_service_id(self): user = self.create_default_user() service = generate_sample_model_tree('Service', self.session) self.set_blue_team(user.team) service.team = user.team self.session.add(service) self.session.commit() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 200 def test_unauthorized_service_id(self): self.create_default_user() resp = self.auth_and_get_path('/service/1') assert resp.status_code == 302 ...
f31ab02d9a409e31acf339db2b950216472b8e9e
salesforce/backend/operations.py
salesforce/backend/operations.py
import re from django.db.backends import BaseDatabaseOperations """ Default database operations, with unquoted names. """ class DatabaseOperations(BaseDatabaseOperations): compiler_module = "salesforce.backend.compiler" def __init__(self, connection): # not calling superclass constructor to maintain Django 1.3 support self.connection = connection self._cache = None def connection_init(self): pass def sql_flush(self, style, tables, sequences): return [] def quote_name(self, name): return name def value_to_db_datetime(self, value): """ We let the JSON serializer handle dates for us. """ return value def last_insert_id(self, cursor, db_table, db_column): return cursor.lastrowid
import re from django.db.backends import BaseDatabaseOperations """ Default database operations, with unquoted names. """ class DatabaseOperations(BaseDatabaseOperations): compiler_module = "salesforce.backend.compiler" def __init__(self, connection): # not calling superclass constructor to maintain Django 1.3 support self.connection = connection self._cache = None def connection_init(self): pass def sql_flush(self, style, tables, sequences): return [] def quote_name(self, name): return name def value_to_db_datetime(self, value): """ We let the JSON serializer handle dates for us. """ return value def value_to_db_date(self, value): """ We let the JSON serializer handle dates for us. """ return value def last_insert_id(self, cursor, db_table, db_column): return cursor.lastrowid
Fix bug with Date fields and SOQL.
Fix bug with Date fields and SOQL. Fixes https://github.com/freelancersunion/django-salesforce/issues/10
Python
mit
django-salesforce/django-salesforce,chromakey/django-salesforce,philchristensen/django-salesforce,hynekcer/django-salesforce,chromakey/django-salesforce,hynekcer/django-salesforce,hynekcer/django-salesforce,chromakey/django-salesforce,django-salesforce/django-salesforce,philchristensen/django-salesforce,django-salesforce/django-salesforce,philchristensen/django-salesforce
import re from django.db.backends import BaseDatabaseOperations """ Default database operations, with unquoted names. """ class DatabaseOperations(BaseDatabaseOperations): compiler_module = "salesforce.backend.compiler" def __init__(self, connection): # not calling superclass constructor to maintain Django 1.3 support self.connection = connection self._cache = None def connection_init(self): pass def sql_flush(self, style, tables, sequences): return [] def quote_name(self, name): return name def value_to_db_datetime(self, value): """ We let the JSON serializer handle dates for us. """ return value + + def value_to_db_date(self, value): + """ + We let the JSON serializer handle dates for us. + """ + return value def last_insert_id(self, cursor, db_table, db_column): return cursor.lastrowid +
Fix bug with Date fields and SOQL.
## Code Before: import re from django.db.backends import BaseDatabaseOperations """ Default database operations, with unquoted names. """ class DatabaseOperations(BaseDatabaseOperations): compiler_module = "salesforce.backend.compiler" def __init__(self, connection): # not calling superclass constructor to maintain Django 1.3 support self.connection = connection self._cache = None def connection_init(self): pass def sql_flush(self, style, tables, sequences): return [] def quote_name(self, name): return name def value_to_db_datetime(self, value): """ We let the JSON serializer handle dates for us. """ return value def last_insert_id(self, cursor, db_table, db_column): return cursor.lastrowid ## Instruction: Fix bug with Date fields and SOQL. ## Code After: import re from django.db.backends import BaseDatabaseOperations """ Default database operations, with unquoted names. """ class DatabaseOperations(BaseDatabaseOperations): compiler_module = "salesforce.backend.compiler" def __init__(self, connection): # not calling superclass constructor to maintain Django 1.3 support self.connection = connection self._cache = None def connection_init(self): pass def sql_flush(self, style, tables, sequences): return [] def quote_name(self, name): return name def value_to_db_datetime(self, value): """ We let the JSON serializer handle dates for us. """ return value def value_to_db_date(self, value): """ We let the JSON serializer handle dates for us. """ return value def last_insert_id(self, cursor, db_table, db_column): return cursor.lastrowid
# ... existing code ... """ return value def value_to_db_date(self, value): """ We let the JSON serializer handle dates for us. """ return value def last_insert_id(self, cursor, db_table, db_column): # ... rest of the code ...
10e307a0dda94a9b38a1b7e143ef141e6062566b
skan/pipe.py
skan/pipe.py
from . import pre, csr import imageio import tqdm import numpy as np from skimage import morphology import pandas as pd def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): image_format = (None if self.image_format.get() == 'auto' else self.image_format.get()) results = [] from skan import pre, csr for file in tqdm(filenames): image = imageio.imread(file, format=image_format) if self.scale_metadata_path is not None: md_path = self.scale_metadata_path.get().split(sep=',') meta = image.meta for key in md_path: meta = meta[key] scale = float(meta) else: scale = 1 # measurements will be in pixel units pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() / scale)) pixel_smoothing_radius = (self.smooth_radius.get() * pixel_threshold_radius) thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, offset=self.brightness_offset.get()) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) framedata['squiggle'] = np.log2(framedata['branch-distance'] / framedata['euclidean-distance']) framedata['filename'] = [file] * len(framedata) results.append(framedata) results = pd.concat(results)
from . import pre, csr import imageio import tqdm import numpy as np from skimage import morphology import pandas as pd def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): image_format = None if image_format == 'auto' else image_format results = [] for file in tqdm(filenames): image = imageio.imread(file, format=image_format) if scale_metadata_path is not None: md_path = scale_metadata_path.split(sep=',') meta = image.meta for key in md_path: meta = meta[key] scale = float(meta) else: scale = 1 # measurements will be in pixel units pixel_threshold_radius = int(np.ceil(threshold_radius / scale)) pixel_smoothing_radius = smooth_radius * pixel_threshold_radius thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, offset=brightness_offset) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) framedata['squiggle'] = np.log2(framedata['branch-distance'] / framedata['euclidean-distance']) framedata['filename'] = [file] * len(framedata) results.append(framedata) return pd.concat(results)
Add module for start-to-finish functions
Add module for start-to-finish functions
Python
bsd-3-clause
jni/skan
from . import pre, csr import imageio import tqdm import numpy as np from skimage import morphology import pandas as pd def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): - image_format = (None if self.image_format.get() == 'auto' + image_format = None if image_format == 'auto' else image_format - else self.image_format.get()) results = [] - from skan import pre, csr for file in tqdm(filenames): image = imageio.imread(file, format=image_format) - if self.scale_metadata_path is not None: + if scale_metadata_path is not None: - md_path = self.scale_metadata_path.get().split(sep=',') + md_path = scale_metadata_path.split(sep=',') meta = image.meta for key in md_path: meta = meta[key] scale = float(meta) else: scale = 1 # measurements will be in pixel units - pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() / + pixel_threshold_radius = int(np.ceil(threshold_radius / scale)) + pixel_smoothing_radius = smooth_radius * pixel_threshold_radius - scale)) - pixel_smoothing_radius = (self.smooth_radius.get() * - pixel_threshold_radius) thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, - offset=self.brightness_offset.get()) + offset=brightness_offset) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) framedata['squiggle'] = np.log2(framedata['branch-distance'] / framedata['euclidean-distance']) framedata['filename'] = [file] * len(framedata) results.append(framedata) - results = pd.concat(results) + return pd.concat(results)
Add module for start-to-finish functions
## Code Before: from . import pre, csr import imageio import tqdm import numpy as np from skimage import morphology import pandas as pd def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): image_format = (None if self.image_format.get() == 'auto' else self.image_format.get()) results = [] from skan import pre, csr for file in tqdm(filenames): image = imageio.imread(file, format=image_format) if self.scale_metadata_path is not None: md_path = self.scale_metadata_path.get().split(sep=',') meta = image.meta for key in md_path: meta = meta[key] scale = float(meta) else: scale = 1 # measurements will be in pixel units pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() / scale)) pixel_smoothing_radius = (self.smooth_radius.get() * pixel_threshold_radius) thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, offset=self.brightness_offset.get()) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) framedata['squiggle'] = np.log2(framedata['branch-distance'] / framedata['euclidean-distance']) framedata['filename'] = [file] * len(framedata) results.append(framedata) results = pd.concat(results) ## Instruction: Add module for start-to-finish functions ## Code After: from . import pre, csr import imageio import tqdm import numpy as np from skimage import morphology import pandas as pd def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): image_format = None if image_format == 'auto' else image_format results = [] for file in tqdm(filenames): image = imageio.imread(file, format=image_format) if scale_metadata_path is not None: md_path = scale_metadata_path.split(sep=',') meta = image.meta for key in md_path: meta = meta[key] scale = float(meta) else: scale = 1 # measurements will be in pixel units pixel_threshold_radius = int(np.ceil(threshold_radius / scale)) pixel_smoothing_radius = smooth_radius * pixel_threshold_radius thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, offset=brightness_offset) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) framedata['squiggle'] = np.log2(framedata['branch-distance'] / framedata['euclidean-distance']) framedata['filename'] = [file] * len(framedata) results.append(framedata) return pd.concat(results)
// ... existing code ... def process_images(filenames, image_format, threshold_radius, smooth_radius, brightness_offset, scale_metadata_path): image_format = None if image_format == 'auto' else image_format results = [] for file in tqdm(filenames): image = imageio.imread(file, format=image_format) if scale_metadata_path is not None: md_path = scale_metadata_path.split(sep=',') meta = image.meta for key in md_path: // ... modified code ... else: scale = 1 # measurements will be in pixel units pixel_threshold_radius = int(np.ceil(threshold_radius / scale)) pixel_smoothing_radius = smooth_radius * pixel_threshold_radius thresholded = pre.threshold(image, sigma=pixel_smoothing_radius, radius=pixel_threshold_radius, offset=brightness_offset) skeleton = morphology.skeletonize(thresholded) framedata = csr.summarise(skeleton, spacing=scale) ... framedata['filename'] = [file] * len(framedata) results.append(framedata) return pd.concat(results) // ... rest of the code ...
b56712563e4205ccbf8b98deace4197e2f250361
movement.py
movement.py
if __name__ == "__main__": x, y = 0, 0 steps = 0 while True: dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} if dir in directions: print("You moved %s, " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) break else: print("You tried to move %s but that is not possible." % dir)
if __name__ == "__main__": x, y = 0, 0 steps = 0 while True: dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} abb_directions = {'n': (0, 1), 's' : (0, -1), 'e' : (1, 0), 'w' : (-1, 0)} long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'} dir = dir.lower().replace(" ", "") if dir in directions: print("You moved %s. " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 elif dir in abb_directions: print("You moved %s. " % long_directions[dir]) x += abb_directions[dir][0] y += abb_directions[dir][1] steps += 1 elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) break else: print("You tried to move %s but that is not possible." % dir)
Add abbreviations and space handling
Add abbreviations and space handling
Python
mit
mewturn/Python
if __name__ == "__main__": x, y = 0, 0 steps = 0 while True: dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) - directions = { 'north': (0, 1), + directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} - + + abb_directions = {'n': (0, 1), + 's' : (0, -1), + 'e' : (1, 0), + 'w' : (-1, 0)} + + long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'} + + dir = dir.lower().replace(" ", "") if dir in directions: - print("You moved %s, " % dir) + print("You moved %s. " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 - + + elif dir in abb_directions: + print("You moved %s. " % long_directions[dir]) + x += abb_directions[dir][0] + y += abb_directions[dir][1] + steps += 1 + elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) break else: print("You tried to move %s but that is not possible." % dir)
Add abbreviations and space handling
## Code Before: if __name__ == "__main__": x, y = 0, 0 steps = 0 while True: dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} if dir in directions: print("You moved %s, " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) break else: print("You tried to move %s but that is not possible." % dir) ## Instruction: Add abbreviations and space handling ## Code After: if __name__ == "__main__": x, y = 0, 0 steps = 0 while True: dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} abb_directions = {'n': (0, 1), 's' : (0, -1), 'e' : (1, 0), 'w' : (-1, 0)} long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'} dir = dir.lower().replace(" ", "") if dir in directions: print("You moved %s. " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 elif dir in abb_directions: print("You moved %s. " % long_directions[dir]) x += abb_directions[dir][0] y += abb_directions[dir][1] steps += 1 elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) break else: print("You tried to move %s but that is not possible." % dir)
// ... existing code ... dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y))) directions = { 'north': (0, 1), 'south' : (0, -1), 'east' : (1, 0), 'west' : (-1, 0)} abb_directions = {'n': (0, 1), 's' : (0, -1), 'e' : (1, 0), 'w' : (-1, 0)} long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'} dir = dir.lower().replace(" ", "") if dir in directions: print("You moved %s. " % dir) x += directions[dir][0] y += directions[dir][1] steps += 1 elif dir in abb_directions: print("You moved %s. " % long_directions[dir]) x += abb_directions[dir][0] y += abb_directions[dir][1] steps += 1 elif dir == "leave": print("Goodbye. You moved %s steps in total." % steps) // ... rest of the code ...
a0aa74d9e6295e34f02b4eefd76e7eb9a1e6425f
node/floor_divide.py
node/floor_divide.py
from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))]
from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))] @Node.test_func(["134", 1], [["134"]]) @Node.test_func(["1234", 2], [["12", "34"]]) @Node.test_func(["1234", 3], [["1", "2", "34"]]) @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]]) def chunk(self, inp:Node.indexable, num:Node.number): """Return inp seperated into num groups""" rtn = [] last = 0 size = len(inp)//num for i in range(size, len(inp), size): rtn.append(inp[last:i]) last = i if len(rtn) != num: rtn.append(inp[last:]) else: rtn[-1] += inp[last:] if len(rtn): if isinstance(inp, str): rtn[-1] = "".join(rtn[-1]) else: rtn[-1] = type(inp)(rtn[-1]) return [rtn]
Add a group chunk, chunks a list into N groups
Add a group chunk, chunks a list into N groups
Python
mit
muddyfish/PYKE,muddyfish/PYKE
from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))] + + @Node.test_func(["134", 1], [["134"]]) + @Node.test_func(["1234", 2], [["12", "34"]]) + @Node.test_func(["1234", 3], [["1", "2", "34"]]) + @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]]) + def chunk(self, inp:Node.indexable, num:Node.number): + """Return inp seperated into num groups""" + rtn = [] + last = 0 + size = len(inp)//num + for i in range(size, len(inp), size): + rtn.append(inp[last:i]) + last = i + if len(rtn) != num: + rtn.append(inp[last:]) + else: + rtn[-1] += inp[last:] + if len(rtn): + if isinstance(inp, str): + rtn[-1] = "".join(rtn[-1]) + else: + rtn[-1] = type(inp)(rtn[-1]) + return [rtn]
Add a group chunk, chunks a list into N groups
## Code Before: from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))] ## Instruction: Add a group chunk, chunks a list into N groups ## Code After: from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))] @Node.test_func(["134", 1], [["134"]]) @Node.test_func(["1234", 2], [["12", "34"]]) @Node.test_func(["1234", 3], [["1", "2", "34"]]) @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]]) def chunk(self, inp:Node.indexable, num:Node.number): """Return inp seperated into num groups""" rtn = [] last = 0 size = len(inp)//num for i in range(size, len(inp), size): rtn.append(inp[last:i]) last = i if len(rtn) != num: rtn.append(inp[last:]) else: rtn[-1] += inp[last:] if len(rtn): if isinstance(inp, str): rtn[-1] = "".join(rtn[-1]) else: rtn[-1] = type(inp)(rtn[-1]) return [rtn]
// ... existing code ... followed by two empty strings.""" return [list(string.partition(sep))] @Node.test_func(["134", 1], [["134"]]) @Node.test_func(["1234", 2], [["12", "34"]]) @Node.test_func(["1234", 3], [["1", "2", "34"]]) @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]]) def chunk(self, inp:Node.indexable, num:Node.number): """Return inp seperated into num groups""" rtn = [] last = 0 size = len(inp)//num for i in range(size, len(inp), size): rtn.append(inp[last:i]) last = i if len(rtn) != num: rtn.append(inp[last:]) else: rtn[-1] += inp[last:] if len(rtn): if isinstance(inp, str): rtn[-1] = "".join(rtn[-1]) else: rtn[-1] = type(inp)(rtn[-1]) return [rtn] // ... rest of the code ...
f353ee5d2e2cf5fd4ee86776fc7e5ee6cb8a3238
sierra_adapter/build_windows.py
sierra_adapter/build_windows.py
import datetime as dt import json import boto3 import docopt import maya args = docopt.docopt(__doc__) start = maya.parse(args['--start']).datetime() end = maya.parse(args['--end']).datetime() minutes = int(args['--interval'] or 30) resource = args['--resource'] assert resource in ('bibs', 'items') def generate_windows(start, end, minutes): current = start while current <= end: yield { 'start': current.isoformat(), 'end': (current + dt.timedelta(minutes=minutes)).isoformat(), } current += dt.timedelta(minutes=minutes - 1) client = boto3.client('sns') for window in generate_windows(start, end, minutes): resp = client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), Subject=f'Window sent by {__file__}' ) print(resp)
import datetime as dt import json import math import boto3 import docopt import maya import tqdm args = docopt.docopt(__doc__) start = maya.parse(args['--start']).datetime() end = maya.parse(args['--end']).datetime() minutes = int(args['--interval'] or 30) resource = args['--resource'] assert resource in ('bibs', 'items') def generate_windows(start, end, minutes): current = start while current <= end: yield { 'start': current.isoformat(), 'end': (current + dt.timedelta(minutes=minutes)).isoformat(), } current += dt.timedelta(minutes=minutes - 1) client = boto3.client('sns') for window in tqdm.tqdm( generate_windows(start, end, minutes), total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1)) ): client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), Subject=f'Window sent by {__file__}' )
Print a progress meter when pushing windows
Print a progress meter when pushing windows
Python
mit
wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api
import datetime as dt import json + import math import boto3 import docopt import maya + import tqdm args = docopt.docopt(__doc__) start = maya.parse(args['--start']).datetime() end = maya.parse(args['--end']).datetime() minutes = int(args['--interval'] or 30) resource = args['--resource'] assert resource in ('bibs', 'items') def generate_windows(start, end, minutes): current = start while current <= end: yield { 'start': current.isoformat(), 'end': (current + dt.timedelta(minutes=minutes)).isoformat(), } current += dt.timedelta(minutes=minutes - 1) client = boto3.client('sns') + for window in tqdm.tqdm( - for window in generate_windows(start, end, minutes): + generate_windows(start, end, minutes), + total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1)) + ): - resp = client.publish( + client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), Subject=f'Window sent by {__file__}' ) - print(resp)
Print a progress meter when pushing windows
## Code Before: import datetime as dt import json import boto3 import docopt import maya args = docopt.docopt(__doc__) start = maya.parse(args['--start']).datetime() end = maya.parse(args['--end']).datetime() minutes = int(args['--interval'] or 30) resource = args['--resource'] assert resource in ('bibs', 'items') def generate_windows(start, end, minutes): current = start while current <= end: yield { 'start': current.isoformat(), 'end': (current + dt.timedelta(minutes=minutes)).isoformat(), } current += dt.timedelta(minutes=minutes - 1) client = boto3.client('sns') for window in generate_windows(start, end, minutes): resp = client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), Subject=f'Window sent by {__file__}' ) print(resp) ## Instruction: Print a progress meter when pushing windows ## Code After: import datetime as dt import json import math import boto3 import docopt import maya import tqdm args = docopt.docopt(__doc__) start = maya.parse(args['--start']).datetime() end = maya.parse(args['--end']).datetime() minutes = int(args['--interval'] or 30) resource = args['--resource'] assert resource in ('bibs', 'items') def generate_windows(start, end, minutes): current = start while current <= end: yield { 'start': current.isoformat(), 'end': (current + dt.timedelta(minutes=minutes)).isoformat(), } current += dt.timedelta(minutes=minutes - 1) client = boto3.client('sns') for window in tqdm.tqdm( generate_windows(start, end, minutes), total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1)) ): client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), Subject=f'Window sent by {__file__}' )
// ... existing code ... import datetime as dt import json import math import boto3 // ... modified code ... import docopt import maya import tqdm args = docopt.docopt(__doc__) ... client = boto3.client('sns') for window in tqdm.tqdm( generate_windows(start, end, minutes), total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1)) ): client.publish( TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows', Message=json.dumps(window), ... Subject=f'Window sent by {__file__}' ) // ... rest of the code ...
b419e78a42e7b8f073bc5d9502dffc97c5d627fb
apps/chats/forms.py
apps/chats/forms.py
from django import forms from django.contrib.auth.models import User from chats.models import Chat from profiles.models import FriendGroup class PublicChatForm(forms.ModelForm): """Public-facing Chat form used in the web-interface for users.""" class Meta: fields = ('text',) model = Chat
from django import forms from django.contrib.auth.models import User from chats.models import Chat from profiles.models import FriendGroup class PublicChatForm(forms.ModelForm): """Public-facing Chat form used in the web-interface for users.""" class Meta: fields = ( 'friend_groups', 'text', ) model = Chat
Add friend_groups to the ChatForm
Add friend_groups to the ChatForm
Python
mit
tofumatt/quotes,tofumatt/quotes
from django import forms from django.contrib.auth.models import User from chats.models import Chat from profiles.models import FriendGroup class PublicChatForm(forms.ModelForm): """Public-facing Chat form used in the web-interface for users.""" class Meta: - fields = ('text',) + fields = ( + 'friend_groups', + 'text', + ) model = Chat
Add friend_groups to the ChatForm
## Code Before: from django import forms from django.contrib.auth.models import User from chats.models import Chat from profiles.models import FriendGroup class PublicChatForm(forms.ModelForm): """Public-facing Chat form used in the web-interface for users.""" class Meta: fields = ('text',) model = Chat ## Instruction: Add friend_groups to the ChatForm ## Code After: from django import forms from django.contrib.auth.models import User from chats.models import Chat from profiles.models import FriendGroup class PublicChatForm(forms.ModelForm): """Public-facing Chat form used in the web-interface for users.""" class Meta: fields = ( 'friend_groups', 'text', ) model = Chat
... class Meta: fields = ( 'friend_groups', 'text', ) model = Chat ...
27b9bd22bb43b8b86ae1c40a90c1fae7157dcb86
app/tests.py
app/tests.py
from app.test_base import BaseTestCase class TestTopLevelFunctions(BaseTestCase): def test_index_response(self): response = self.client.get('/') self.assert200(response)
from app.test_base import BaseTestCase class TestTopLevelFunctions(BaseTestCase): def test_index_response(self): response = self.client.get('/') self.assert200(response) def test_login_required(self): self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd') self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew') self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew') self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew') self.check_login_required('/settings', '/login?next=%2Fsettings') self.check_login_required('/review', '/login?next=%2Freview') self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew') self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs') def check_login_required(self, attempted_location, redirected_location): response = self.client.get(attempted_location) self.assertTrue(response.status_code in (301, 302)) self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
Add test to verify login required for protected pages
Add test to verify login required for protected pages
Python
mit
rtfoley/scorepy,rtfoley/scorepy,rtfoley/scorepy
from app.test_base import BaseTestCase class TestTopLevelFunctions(BaseTestCase): def test_index_response(self): response = self.client.get('/') self.assert200(response) + def test_login_required(self): + self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd') + self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew') + self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew') + self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew') + self.check_login_required('/settings', '/login?next=%2Fsettings') + self.check_login_required('/review', '/login?next=%2Freview') + self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew') + self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs') + + def check_login_required(self, attempted_location, redirected_location): + response = self.client.get(attempted_location) + self.assertTrue(response.status_code in (301, 302)) + self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location) +
Add test to verify login required for protected pages
## Code Before: from app.test_base import BaseTestCase class TestTopLevelFunctions(BaseTestCase): def test_index_response(self): response = self.client.get('/') self.assert200(response) ## Instruction: Add test to verify login required for protected pages ## Code After: from app.test_base import BaseTestCase class TestTopLevelFunctions(BaseTestCase): def test_index_response(self): response = self.client.get('/') self.assert200(response) def test_login_required(self): self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd') self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew') self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew') self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew') self.check_login_required('/settings', '/login?next=%2Fsettings') self.check_login_required('/review', '/login?next=%2Freview') self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew') self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs') def check_login_required(self, attempted_location, redirected_location): response = self.client.get(attempted_location) self.assertTrue(response.status_code in (301, 302)) self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
// ... existing code ... response = self.client.get('/') self.assert200(response) def test_login_required(self): self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd') self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew') self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew') self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew') self.check_login_required('/settings', '/login?next=%2Fsettings') self.check_login_required('/review', '/login?next=%2Freview') self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew') self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs') def check_login_required(self, attempted_location, redirected_location): response = self.client.get(attempted_location) self.assertTrue(response.status_code in (301, 302)) self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location) // ... rest of the code ...
c6298a573dc3188b8c57954287d78e7da253483a
lot/urls.py
lot/urls.py
from django.conf.urls import patterns, url from . import views urlpatterns = patterns("", url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), )
from django.conf.urls import url from . import views urlpatterns = [ url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), ]
Update to new-style urlpatterns format
Update to new-style urlpatterns format
Python
bsd-3-clause
ABASystems/django-lot
- from django.conf.urls import patterns, url + from django.conf.urls import url from . import views - urlpatterns = patterns("", + urlpatterns = [ url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), - ) + ]
Update to new-style urlpatterns format
## Code Before: from django.conf.urls import patterns, url from . import views urlpatterns = patterns("", url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), ) ## Instruction: Update to new-style urlpatterns format ## Code After: from django.conf.urls import url from . import views urlpatterns = [ url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), ]
// ... existing code ... from django.conf.urls import url from . import views urlpatterns = [ url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"), ] // ... rest of the code ...
9b10f600b5611380f72fe2aeacfe2ee6f02e4e3a
kicad_footprint_load.py
kicad_footprint_load.py
import pcbnew import sys import os pretties = [] for dirname, dirnames, filenames in os.walk(sys.argv[1]): # don't go into any .git directories. if '.git' in dirnames: dirnames.remove('.git') for filename in filenames: if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'): pretties.append(os.path.realpath(dirname)) break src_plugin = pcbnew.IO_MGR.PluginFind(1) for libpath in pretties: #Ignore paths with unicode as KiCad can't deal with them in enumerate list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
import pcbnew import sys import os pretties = [] for dirname, dirnames, filenames in os.walk(sys.argv[1]): # don't go into any .git directories. if '.git' in dirnames: dirnames.remove('.git') for filename in filenames: if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'): pretties.append(os.path.realpath(dirname)) break src_plugin = pcbnew.IO_MGR.PluginFind(1) for libpath in pretties: list_of_footprints = src_plugin.FootprintEnumerate(libpath)
Switch to old invocation of FootprintEnumerate
Switch to old invocation of FootprintEnumerate
Python
mit
monostable/haskell-kicad-data,monostable/haskell-kicad-data,kasbah/haskell-kicad-data
import pcbnew import sys import os pretties = [] for dirname, dirnames, filenames in os.walk(sys.argv[1]): # don't go into any .git directories. if '.git' in dirnames: dirnames.remove('.git') for filename in filenames: if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'): pretties.append(os.path.realpath(dirname)) break src_plugin = pcbnew.IO_MGR.PluginFind(1) for libpath in pretties: - #Ignore paths with unicode as KiCad can't deal with them in enumerate - list_of_footprints = src_plugin.FootprintEnumerate(libpath, False) + list_of_footprints = src_plugin.FootprintEnumerate(libpath)
Switch to old invocation of FootprintEnumerate
## Code Before: import pcbnew import sys import os pretties = [] for dirname, dirnames, filenames in os.walk(sys.argv[1]): # don't go into any .git directories. if '.git' in dirnames: dirnames.remove('.git') for filename in filenames: if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'): pretties.append(os.path.realpath(dirname)) break src_plugin = pcbnew.IO_MGR.PluginFind(1) for libpath in pretties: #Ignore paths with unicode as KiCad can't deal with them in enumerate list_of_footprints = src_plugin.FootprintEnumerate(libpath, False) ## Instruction: Switch to old invocation of FootprintEnumerate ## Code After: import pcbnew import sys import os pretties = [] for dirname, dirnames, filenames in os.walk(sys.argv[1]): # don't go into any .git directories. if '.git' in dirnames: dirnames.remove('.git') for filename in filenames: if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'): pretties.append(os.path.realpath(dirname)) break src_plugin = pcbnew.IO_MGR.PluginFind(1) for libpath in pretties: list_of_footprints = src_plugin.FootprintEnumerate(libpath)
# ... existing code ... for libpath in pretties: list_of_footprints = src_plugin.FootprintEnumerate(libpath) # ... rest of the code ...
32c40710a562b194385f2340bf882cb3709b74e3
masquerade/urls.py
masquerade/urls.py
from django.conf.urls import patterns, url urlpatterns = patterns('', url(r'^mask/$', 'masquerade.views.mask'), url(r'^unmask/$', 'masquerade.views.unmask'), )
from django.conf.urls import patterns, url from masquerade.views import mask from masquerade.views import unmask urlpatterns = [ url(r'^mask/$', mask), url(r'^unmask/$', unmask), ]
Fix Django 1.10 deprecation warning
Fix Django 1.10 deprecation warning
Python
apache-2.0
erikcw/django-masquerade,erikcw/django-masquerade,erikcw/django-masquerade
from django.conf.urls import patterns, url + from masquerade.views import mask + from masquerade.views import unmask - urlpatterns = patterns('', - url(r'^mask/$', 'masquerade.views.mask'), - url(r'^unmask/$', 'masquerade.views.unmask'), - ) + + urlpatterns = [ + url(r'^mask/$', mask), + url(r'^unmask/$', unmask), + ] +
Fix Django 1.10 deprecation warning
## Code Before: from django.conf.urls import patterns, url urlpatterns = patterns('', url(r'^mask/$', 'masquerade.views.mask'), url(r'^unmask/$', 'masquerade.views.unmask'), ) ## Instruction: Fix Django 1.10 deprecation warning ## Code After: from django.conf.urls import patterns, url from masquerade.views import mask from masquerade.views import unmask urlpatterns = [ url(r'^mask/$', mask), url(r'^unmask/$', unmask), ]
// ... existing code ... from django.conf.urls import patterns, url from masquerade.views import mask from masquerade.views import unmask urlpatterns = [ url(r'^mask/$', mask), url(r'^unmask/$', unmask), ] // ... rest of the code ...
07455e5821d21c988c7c5fcda9345e99355eb4e7
redash/__init__.py
redash/__init__.py
import json import urlparse from flask import Flask, make_response from flask.ext.restful import Api from flask_peewee.db import Database import redis from redash import settings, utils __version__ = '0.3.2' app = Flask(__name__, template_folder=settings.STATIC_ASSETS_PATH, static_folder=settings.STATIC_ASSETS_PATH, static_path='/static') api = Api(app) # configure our database settings.DATABASE_CONFIG.update({'threadlocals': True}) app.config['DATABASE'] = settings.DATABASE_CONFIG db = Database(app) from redash.authentication import setup_authentication auth = setup_authentication(app) @api.representation('application/json') def json_representation(data, code, headers=None): resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code) resp.headers.extend(headers or {}) return resp redis_url = urlparse.urlparse(settings.REDIS_URL) redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password) from redash import data data_manager = data.Manager(redis_connection, db) from redash import controllers
import json import urlparse from flask import Flask, make_response from flask.ext.restful import Api from flask_peewee.db import Database import redis from redash import settings, utils __version__ = '0.3.2' app = Flask(__name__, template_folder=settings.STATIC_ASSETS_PATH, static_folder=settings.STATIC_ASSETS_PATH, static_path='/static') api = Api(app) # configure our database settings.DATABASE_CONFIG.update({'threadlocals': True}) app.config['DATABASE'] = settings.DATABASE_CONFIG db = Database(app) from redash.authentication import setup_authentication auth = setup_authentication(app) @api.representation('application/json') def json_representation(data, code, headers=None): resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code) resp.headers.extend(headers or {}) return resp redis_url = urlparse.urlparse(settings.REDIS_URL) if redis_url.path: redis_db = redis_url.path[1] else: redis_db = 0 redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password) from redash import data data_manager = data.Manager(redis_connection, db) from redash import controllers
Use database number from redis url if available.
Use database number from redis url if available.
Python
bsd-2-clause
chriszs/redash,imsally/redash,44px/redash,guaguadev/redash,denisov-vlad/redash,rockwotj/redash,44px/redash,rockwotj/redash,getredash/redash,ninneko/redash,akariv/redash,amino-data/redash,akariv/redash,imsally/redash,EverlyWell/redash,getredash/redash,easytaxibr/redash,M32Media/redash,vishesh92/redash,ninneko/redash,getredash/redash,easytaxibr/redash,crowdworks/redash,hudl/redash,ninneko/redash,denisov-vlad/redash,stefanseifert/redash,amino-data/redash,useabode/redash,guaguadev/redash,jmvasquez/redashtest,alexanderlz/redash,moritz9/redash,pubnative/redash,denisov-vlad/redash,imsally/redash,rockwotj/redash,pubnative/redash,vishesh92/redash,useabode/redash,crowdworks/redash,akariv/redash,44px/redash,easytaxibr/redash,getredash/redash,chriszs/redash,akariv/redash,guaguadev/redash,44px/redash,stefanseifert/redash,denisov-vlad/redash,amino-data/redash,moritz9/redash,M32Media/redash,crowdworks/redash,M32Media/redash,M32Media/redash,alexanderlz/redash,crowdworks/redash,jmvasquez/redashtest,stefanseifert/redash,denisov-vlad/redash,pubnative/redash,EverlyWell/redash,moritz9/redash,stefanseifert/redash,stefanseifert/redash,pubnative/redash,imsally/redash,EverlyWell/redash,akariv/redash,ninneko/redash,moritz9/redash,jmvasquez/redashtest,useabode/redash,vishesh92/redash,chriszs/redash,getredash/redash,jmvasquez/redashtest,useabode/redash,easytaxibr/redash,ninneko/redash,amino-data/redash,pubnative/redash,hudl/redash,EverlyWell/redash,guaguadev/redash,jmvasquez/redashtest,easytaxibr/redash,hudl/redash,vishesh92/redash,chriszs/redash,rockwotj/redash,alexanderlz/redash,hudl/redash,alexanderlz/redash,guaguadev/redash
import json import urlparse from flask import Flask, make_response from flask.ext.restful import Api from flask_peewee.db import Database import redis from redash import settings, utils __version__ = '0.3.2' app = Flask(__name__, template_folder=settings.STATIC_ASSETS_PATH, static_folder=settings.STATIC_ASSETS_PATH, static_path='/static') api = Api(app) # configure our database settings.DATABASE_CONFIG.update({'threadlocals': True}) app.config['DATABASE'] = settings.DATABASE_CONFIG db = Database(app) from redash.authentication import setup_authentication auth = setup_authentication(app) @api.representation('application/json') def json_representation(data, code, headers=None): resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code) resp.headers.extend(headers or {}) return resp redis_url = urlparse.urlparse(settings.REDIS_URL) + if redis_url.path: + redis_db = redis_url.path[1] + else: + redis_db = 0 - redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password) + redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password) from redash import data data_manager = data.Manager(redis_connection, db) from redash import controllers
Use database number from redis url if available.
## Code Before: import json import urlparse from flask import Flask, make_response from flask.ext.restful import Api from flask_peewee.db import Database import redis from redash import settings, utils __version__ = '0.3.2' app = Flask(__name__, template_folder=settings.STATIC_ASSETS_PATH, static_folder=settings.STATIC_ASSETS_PATH, static_path='/static') api = Api(app) # configure our database settings.DATABASE_CONFIG.update({'threadlocals': True}) app.config['DATABASE'] = settings.DATABASE_CONFIG db = Database(app) from redash.authentication import setup_authentication auth = setup_authentication(app) @api.representation('application/json') def json_representation(data, code, headers=None): resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code) resp.headers.extend(headers or {}) return resp redis_url = urlparse.urlparse(settings.REDIS_URL) redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password) from redash import data data_manager = data.Manager(redis_connection, db) from redash import controllers ## Instruction: Use database number from redis url if available. ## Code After: import json import urlparse from flask import Flask, make_response from flask.ext.restful import Api from flask_peewee.db import Database import redis from redash import settings, utils __version__ = '0.3.2' app = Flask(__name__, template_folder=settings.STATIC_ASSETS_PATH, static_folder=settings.STATIC_ASSETS_PATH, static_path='/static') api = Api(app) # configure our database settings.DATABASE_CONFIG.update({'threadlocals': True}) app.config['DATABASE'] = settings.DATABASE_CONFIG db = Database(app) from redash.authentication import setup_authentication auth = setup_authentication(app) @api.representation('application/json') def json_representation(data, code, headers=None): resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code) resp.headers.extend(headers or {}) return resp redis_url = urlparse.urlparse(settings.REDIS_URL) if redis_url.path: redis_db = redis_url.path[1] else: redis_db = 0 redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password) from redash import data data_manager = data.Manager(redis_connection, db) from redash import controllers
// ... existing code ... redis_url = urlparse.urlparse(settings.REDIS_URL) if redis_url.path: redis_db = redis_url.path[1] else: redis_db = 0 redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password) from redash import data // ... rest of the code ...
c4e1f1c147783a4a735dd943d5d7491302de300e
csunplugged/config/urls.py
csunplugged/config/urls.py
from django.conf.urls import include, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.conf import settings from django.conf.urls.static import static urlpatterns = i18n_patterns( url(r'', include('general.urls', namespace='general')), url(r'^topics/', include('topics.urls', namespace='topics')), url(r'^resources/', include('resources.urls', namespace='resources')), url(r'^admin/', include(admin.site.urls)), ) # ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT)
from django.conf.urls import include, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin urlpatterns = i18n_patterns( url(r'', include('general.urls', namespace='general')), url(r'^topics/', include('topics.urls', namespace='topics')), url(r'^resources/', include('resources.urls', namespace='resources')), url(r'^admin/', include(admin.site.urls)), )
Remove unused static URL pathing
Remove unused static URL pathing
Python
mit
uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged
from django.conf.urls import include, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin - from django.conf import settings - from django.conf.urls.static import static urlpatterns = i18n_patterns( url(r'', include('general.urls', namespace='general')), url(r'^topics/', include('topics.urls', namespace='topics')), url(r'^resources/', include('resources.urls', namespace='resources')), url(r'^admin/', include(admin.site.urls)), ) - # ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT)
Remove unused static URL pathing
## Code Before: from django.conf.urls import include, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.conf import settings from django.conf.urls.static import static urlpatterns = i18n_patterns( url(r'', include('general.urls', namespace='general')), url(r'^topics/', include('topics.urls', namespace='topics')), url(r'^resources/', include('resources.urls', namespace='resources')), url(r'^admin/', include(admin.site.urls)), ) # ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT) ## Instruction: Remove unused static URL pathing ## Code After: from django.conf.urls import include, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin urlpatterns = i18n_patterns( url(r'', include('general.urls', namespace='general')), url(r'^topics/', include('topics.urls', namespace='topics')), url(r'^resources/', include('resources.urls', namespace='resources')), url(r'^admin/', include(admin.site.urls)), )
... from django.conf.urls.i18n import i18n_patterns from django.contrib import admin urlpatterns = i18n_patterns( ... url(r'^admin/', include(admin.site.urls)), ) ...
252ffda53d494403133fdb1986c92422264406d8
tests_app/tests/unit/serializers/models.py
tests_app/tests/unit/serializers/models.py
import os from django.db import models from django.conf import settings class UserModel(models.Model): name = models.CharField(max_length=20) upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers') class CommentModel(models.Model): user = models.ForeignKey( UserModel, related_name='comments', on_delete=models.CASCADE, ) users_liked = models.ManyToManyField(UserModel, blank=True, null=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) attachment = models.FileField( upload_to=upload_to, blank=True, null=True, max_length=500) hidden_text = models.CharField(max_length=200, blank=True, null=True)
import os from django.db import models from django.conf import settings class UserModel(models.Model): name = models.CharField(max_length=20) upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers') class CommentModel(models.Model): user = models.ForeignKey( UserModel, related_name='comments', on_delete=models.CASCADE, ) users_liked = models.ManyToManyField(UserModel, blank=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) attachment = models.FileField( upload_to=upload_to, blank=True, null=True, max_length=500) hidden_text = models.CharField(max_length=200, blank=True, null=True)
Fix CommentModel m2m null warning
Fix CommentModel m2m null warning
Python
mit
chibisov/drf-extensions
import os from django.db import models from django.conf import settings class UserModel(models.Model): name = models.CharField(max_length=20) upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers') class CommentModel(models.Model): user = models.ForeignKey( UserModel, related_name='comments', on_delete=models.CASCADE, ) - users_liked = models.ManyToManyField(UserModel, blank=True, null=True) + users_liked = models.ManyToManyField(UserModel, blank=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) attachment = models.FileField( upload_to=upload_to, blank=True, null=True, max_length=500) hidden_text = models.CharField(max_length=200, blank=True, null=True)
Fix CommentModel m2m null warning
## Code Before: import os from django.db import models from django.conf import settings class UserModel(models.Model): name = models.CharField(max_length=20) upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers') class CommentModel(models.Model): user = models.ForeignKey( UserModel, related_name='comments', on_delete=models.CASCADE, ) users_liked = models.ManyToManyField(UserModel, blank=True, null=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) attachment = models.FileField( upload_to=upload_to, blank=True, null=True, max_length=500) hidden_text = models.CharField(max_length=200, blank=True, null=True) ## Instruction: Fix CommentModel m2m null warning ## Code After: import os from django.db import models from django.conf import settings class UserModel(models.Model): name = models.CharField(max_length=20) upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers') class CommentModel(models.Model): user = models.ForeignKey( UserModel, related_name='comments', on_delete=models.CASCADE, ) users_liked = models.ManyToManyField(UserModel, blank=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) attachment = models.FileField( upload_to=upload_to, blank=True, null=True, max_length=500) hidden_text = models.CharField(max_length=200, blank=True, null=True)
... on_delete=models.CASCADE, ) users_liked = models.ManyToManyField(UserModel, blank=True) title = models.CharField(max_length=20) text = models.CharField(max_length=200) ...
20147b8b8a80ef8ab202d916bf1cdfb67d4753d3
SelfTests.py
SelfTests.py
import os import unittest from Logger import Logger class TestLogger(unittest.TestCase): def test_file_handling(self): testLog = Logger("testLog") ## Check if program can create and open file self.assertTrue(testLog.opened) returns = testLog.close() ## Check if logger correctly signs bool OPENED and returns ## 0 as succes. self.assertFalse(testLog.opened) self.assertEqual(returns,0) returns = testLog.close() ## Check if logger returns 1 when trying to close already ## closed file self.assertEqual(returns,1) ## Do cleanup: os.remove(testLog.name) def test_logging(self): testLog = Logger("testLog") testLog.save_line("TestLine") testLog.close() logfile = open(testLog.name) content = logfile.read() logfile.close() saved = content.split(" : ") self.assertEqual(saved[1],"TestLine") ## cleanup os.remove(testLog.name) if __name__ == '__main__': unittest.main()
import os import unittest from Logger import Logger class TestLogger(unittest.TestCase): def test_file_handling(self): testLog = Logger("testLog") ## Check if program can create and open file self.assertTrue(testLog.opened) returns = testLog.close() ## Check if logger correctly signs bool OPENED and returns ## 0 as succes. self.assertFalse(testLog.opened) self.assertEqual(returns,0) returns = testLog.close() ## Check if logger returns 1 when trying to close already ## closed file self.assertEqual(returns,1) ## Do cleanup: os.remove(testLog.name) def test_logging(self): testLog = Logger("testLog") testPhrase = "TestLine\r\n" testLog.save_line(testPhrase) testLog.close() logfile = open(testLog.name) content = logfile.read() logfile.close() saved = content.split(" : ") ## Check if saved data corresponds self.assertEqual(saved[1],testPhrase) ## cleanup os.remove(testLog.name) if __name__ == '__main__': unittest.main()
Test of logger is testing an testPhrase instead of two manually writen strings
Test of logger is testing an testPhrase instead of two manually writen strings Signed-off-by: TeaPackCZ <[email protected]>
Python
mit
TeaPackCZ/RobotZed,TeaPackCZ/RobotZed
import os import unittest from Logger import Logger class TestLogger(unittest.TestCase): def test_file_handling(self): testLog = Logger("testLog") ## Check if program can create and open file self.assertTrue(testLog.opened) returns = testLog.close() ## Check if logger correctly signs bool OPENED and returns ## 0 as succes. self.assertFalse(testLog.opened) self.assertEqual(returns,0) returns = testLog.close() ## Check if logger returns 1 when trying to close already ## closed file self.assertEqual(returns,1) ## Do cleanup: os.remove(testLog.name) def test_logging(self): testLog = Logger("testLog") + testPhrase = "TestLine\r\n" - testLog.save_line("TestLine") + testLog.save_line(testPhrase) testLog.close() logfile = open(testLog.name) content = logfile.read() logfile.close() saved = content.split(" : ") + ## Check if saved data corresponds - self.assertEqual(saved[1],"TestLine") + self.assertEqual(saved[1],testPhrase) ## cleanup os.remove(testLog.name) if __name__ == '__main__': unittest.main()
Test of logger is testing an testPhrase instead of two manually writen strings
## Code Before: import os import unittest from Logger import Logger class TestLogger(unittest.TestCase): def test_file_handling(self): testLog = Logger("testLog") ## Check if program can create and open file self.assertTrue(testLog.opened) returns = testLog.close() ## Check if logger correctly signs bool OPENED and returns ## 0 as succes. self.assertFalse(testLog.opened) self.assertEqual(returns,0) returns = testLog.close() ## Check if logger returns 1 when trying to close already ## closed file self.assertEqual(returns,1) ## Do cleanup: os.remove(testLog.name) def test_logging(self): testLog = Logger("testLog") testLog.save_line("TestLine") testLog.close() logfile = open(testLog.name) content = logfile.read() logfile.close() saved = content.split(" : ") self.assertEqual(saved[1],"TestLine") ## cleanup os.remove(testLog.name) if __name__ == '__main__': unittest.main() ## Instruction: Test of logger is testing an testPhrase instead of two manually writen strings ## Code After: import os import unittest from Logger import Logger class TestLogger(unittest.TestCase): def test_file_handling(self): testLog = Logger("testLog") ## Check if program can create and open file self.assertTrue(testLog.opened) returns = testLog.close() ## Check if logger correctly signs bool OPENED and returns ## 0 as succes. self.assertFalse(testLog.opened) self.assertEqual(returns,0) returns = testLog.close() ## Check if logger returns 1 when trying to close already ## closed file self.assertEqual(returns,1) ## Do cleanup: os.remove(testLog.name) def test_logging(self): testLog = Logger("testLog") testPhrase = "TestLine\r\n" testLog.save_line(testPhrase) testLog.close() logfile = open(testLog.name) content = logfile.read() logfile.close() saved = content.split(" : ") ## Check if saved data corresponds self.assertEqual(saved[1],testPhrase) ## cleanup os.remove(testLog.name) if __name__ == '__main__': unittest.main()
# ... existing code ... def test_logging(self): testLog = Logger("testLog") testPhrase = "TestLine\r\n" testLog.save_line(testPhrase) testLog.close() logfile = open(testLog.name) # ... modified code ... logfile.close() saved = content.split(" : ") ## Check if saved data corresponds self.assertEqual(saved[1],testPhrase) ## cleanup os.remove(testLog.name) # ... rest of the code ...
bb0b72333b715956740373c3ba80a8193b99a8cc
app/services/updater_service.py
app/services/updater_service.py
from app.system.updater import check_updates, do_upgrade, run_ansible from app.views import SimpleBackgroundView from .base import BaseService, BlockingServiceStart class UpdaterService(BaseService, BlockingServiceStart): def __init__(self, observer=None): super().__init__(observer=observer) self._view = SimpleBackgroundView("Checking for updates.") def on_service_start(self): values = check_updates() for val in values: self._view.args["subtitle"] = "Working with: " + str(val) do_upgrade([val]) if values: run_ansible() def view(self): return self._view
from app.system.updater import check_updates, do_upgrade, run_ansible from app.views import SimpleBackgroundView from .base import BaseService, BlockingServiceStart class UpdaterService(BaseService, BlockingServiceStart): def __init__(self, observer=None): super().__init__(observer=observer) self._view = SimpleBackgroundView("Checking for updates.") def on_service_start(self): values = check_updates() for val in values: self._view.args["subtitle"] = "Working with: " + str(val) do_upgrade([val]) if values: self._view.args["subtitle"] = "Finishing up..." run_ansible() def view(self): return self._view
Add message before running ansible.
Add message before running ansible.
Python
mit
supersaiyanmode/HomePiServer,supersaiyanmode/HomePiServer,supersaiyanmode/HomePiServer
from app.system.updater import check_updates, do_upgrade, run_ansible from app.views import SimpleBackgroundView from .base import BaseService, BlockingServiceStart class UpdaterService(BaseService, BlockingServiceStart): def __init__(self, observer=None): super().__init__(observer=observer) self._view = SimpleBackgroundView("Checking for updates.") def on_service_start(self): values = check_updates() for val in values: self._view.args["subtitle"] = "Working with: " + str(val) do_upgrade([val]) if values: + self._view.args["subtitle"] = "Finishing up..." run_ansible() def view(self): return self._view
Add message before running ansible.
## Code Before: from app.system.updater import check_updates, do_upgrade, run_ansible from app.views import SimpleBackgroundView from .base import BaseService, BlockingServiceStart class UpdaterService(BaseService, BlockingServiceStart): def __init__(self, observer=None): super().__init__(observer=observer) self._view = SimpleBackgroundView("Checking for updates.") def on_service_start(self): values = check_updates() for val in values: self._view.args["subtitle"] = "Working with: " + str(val) do_upgrade([val]) if values: run_ansible() def view(self): return self._view ## Instruction: Add message before running ansible. ## Code After: from app.system.updater import check_updates, do_upgrade, run_ansible from app.views import SimpleBackgroundView from .base import BaseService, BlockingServiceStart class UpdaterService(BaseService, BlockingServiceStart): def __init__(self, observer=None): super().__init__(observer=observer) self._view = SimpleBackgroundView("Checking for updates.") def on_service_start(self): values = check_updates() for val in values: self._view.args["subtitle"] = "Working with: " + str(val) do_upgrade([val]) if values: self._view.args["subtitle"] = "Finishing up..." run_ansible() def view(self): return self._view
... do_upgrade([val]) if values: self._view.args["subtitle"] = "Finishing up..." run_ansible() ...
866026a5d2f89a8ac76a726720e4fbe812c94eb4
ds/providers/shell.py
ds/providers/shell.py
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command)
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
Fix arg passing to command
Fix arg passing to command
Python
apache-2.0
rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,getsentry/freight,klynton/freight,getsentry/freight,klynton/freight,rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,rshk/freight,jkimbo/freight,jkimbo/freight,rshk/freight,getsentry/freight
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): - command = task.provider_config['command'].format({ + command = task.provider_config['command'].format( - 'environment': task.environment, + environment=task.environment, - 'sha': task.sha, + sha=task.sha, - 'ref': task.ref, + ref=task.ref, - 'task': task.name, + task=task.name, - }) + ) return workspace.run(command)
Fix arg passing to command
## Code Before: from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command) ## Instruction: Fix arg passing to command ## Code After: from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
... def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command) ...
59ec54bbe49013826d2c15ce2162c2e0e335bd57
modules/module_urlsize.py
modules/module_urlsize.py
"""Warns about large files""" def handle_url(bot, user, channel, url, msg): if channel == "#wow": return # inform about large files (over 5MB) size = getUrl(url).getSize() contentType = getUrl(url).getHeaders()['content-type'] if not size: return size = size / 1024 if size > 5: bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
"""Warns about large files""" def handle_url(bot, user, channel, url, msg): if channel == "#wow": return # inform about large files (over 5MB) size = getUrl(url).getSize() headers = getUrl(url).getHeaders()['content-type'] if 'content-type' in headers: contentType = headers['content-type'] else: contentType = "Unknown" if not size: return size = size / 1024 if size > 5: bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
Handle cases where the server doesn't return content-type
Handle cases where the server doesn't return content-type git-svn-id: 056f9092885898c4775d98c479d2d33d00273e45@120 dda364a1-ef19-0410-af65-756c83048fb2
Python
bsd-3-clause
rnyberg/pyfibot,lepinkainen/pyfibot,lepinkainen/pyfibot,rnyberg/pyfibot,EArmour/pyfibot,nigeljonez/newpyfibot,aapa/pyfibot,huqa/pyfibot,huqa/pyfibot,aapa/pyfibot,EArmour/pyfibot
"""Warns about large files""" def handle_url(bot, user, channel, url, msg): if channel == "#wow": return # inform about large files (over 5MB) size = getUrl(url).getSize() - contentType = getUrl(url).getHeaders()['content-type'] + headers = getUrl(url).getHeaders()['content-type'] + if 'content-type' in headers: + contentType = headers['content-type'] + else: + contentType = "Unknown" if not size: return size = size / 1024 if size > 5: bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
Handle cases where the server doesn't return content-type
## Code Before: """Warns about large files""" def handle_url(bot, user, channel, url, msg): if channel == "#wow": return # inform about large files (over 5MB) size = getUrl(url).getSize() contentType = getUrl(url).getHeaders()['content-type'] if not size: return size = size / 1024 if size > 5: bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType)) ## Instruction: Handle cases where the server doesn't return content-type ## Code After: """Warns about large files""" def handle_url(bot, user, channel, url, msg): if channel == "#wow": return # inform about large files (over 5MB) size = getUrl(url).getSize() headers = getUrl(url).getHeaders()['content-type'] if 'content-type' in headers: contentType = headers['content-type'] else: contentType = "Unknown" if not size: return size = size / 1024 if size > 5: bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
# ... existing code ... # inform about large files (over 5MB) size = getUrl(url).getSize() headers = getUrl(url).getHeaders()['content-type'] if 'content-type' in headers: contentType = headers['content-type'] else: contentType = "Unknown" if not size: return # ... rest of the code ...
056d82002c133736a800b08bd071b71c9f5615f8
ci/generate_pipeline_yml.py
ci/generate_pipeline_yml.py
import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
Update TAS versions we test against
Update TAS versions we test against
Python
apache-2.0
cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator
import os from jinja2 import Template - clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] + clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
Update TAS versions we test against
## Code Before: import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml") ## Instruction: Update TAS versions we test against ## Code After: import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
// ... existing code ... from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] // ... rest of the code ...
bbfdbc4b5b6a35105a65910a878be85040cf5263
VMEncryption/main/oscrypto/encryptstates/OSEncryptionState.py
VMEncryption/main/oscrypto/encryptstates/OSEncryptionState.py
from collections import namedtuple class OSEncryptionState(object): def __init__(self, context): super(OSEncryptionState, self).__init__() self.state_executed = False def enter(self): assert 0, "implement enter" def should_exit(self): assert 0, "implement should_exit" OSEncryptionStateContext = namedtuple('OSEncryptionStateContext', ['hutil', 'distro_patcher', 'logger', 'encryption_environment'])
from collections import namedtuple class OSEncryptionState(object): def __init__(self, context): super(OSEncryptionState, self).__init__() def enter(self): assert 0, "implement enter" def should_exit(self): assert 0, "implement should_exit" OSEncryptionStateContext = namedtuple('OSEncryptionStateContext', ['hutil', 'distro_patcher', 'logger', 'encryption_environment'])
Remove var declaration from abstract base class
Remove var declaration from abstract base class
Python
apache-2.0
soumyanishan/azure-linux-extensions,bpramod/azure-linux-extensions,Azure/azure-linux-extensions,Azure/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,Azure/azure-linux-extensions,andyliuliming/azure-linux-extensions,andyliuliming/azure-linux-extensions,jasonzio/azure-linux-extensions,krkhan/azure-linux-extensions,varunkumta/azure-linux-extensions,Azure/azure-linux-extensions,jasonzio/azure-linux-extensions,krkhan/azure-linux-extensions,bpramod/azure-linux-extensions,krkhan/azure-linux-extensions,jasonzio/azure-linux-extensions,soumyanishan/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,soumyanishan/azure-linux-extensions,soumyanishan/azure-linux-extensions,varunkumta/azure-linux-extensions,varunkumta/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,varunkumta/azure-linux-extensions,andyliuliming/azure-linux-extensions,andyliuliming/azure-linux-extensions,bpramod/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,vityagi/azure-linux-extensions,krkhan/azure-linux-extensions,vityagi/azure-linux-extensions,jasonzio/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,bpramod/azure-linux-extensions,soumyanishan/azure-linux-extensions
from collections import namedtuple class OSEncryptionState(object): def __init__(self, context): super(OSEncryptionState, self).__init__() - - self.state_executed = False def enter(self): assert 0, "implement enter" def should_exit(self): assert 0, "implement should_exit" OSEncryptionStateContext = namedtuple('OSEncryptionStateContext', ['hutil', 'distro_patcher', 'logger', 'encryption_environment'])
Remove var declaration from abstract base class
## Code Before: from collections import namedtuple class OSEncryptionState(object): def __init__(self, context): super(OSEncryptionState, self).__init__() self.state_executed = False def enter(self): assert 0, "implement enter" def should_exit(self): assert 0, "implement should_exit" OSEncryptionStateContext = namedtuple('OSEncryptionStateContext', ['hutil', 'distro_patcher', 'logger', 'encryption_environment']) ## Instruction: Remove var declaration from abstract base class ## Code After: from collections import namedtuple class OSEncryptionState(object): def __init__(self, context): super(OSEncryptionState, self).__init__() def enter(self): assert 0, "implement enter" def should_exit(self): assert 0, "implement should_exit" OSEncryptionStateContext = namedtuple('OSEncryptionStateContext', ['hutil', 'distro_patcher', 'logger', 'encryption_environment'])
# ... existing code ... def __init__(self, context): super(OSEncryptionState, self).__init__() def enter(self): # ... rest of the code ...
5b3d38821517f10f9b9da31f28af19e7302de954
dimod/reference/composites/structure.py
dimod/reference/composites/structure.py
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
Update Structure composite to use the new abc
Update Structure composite to use the new abc
Python
apache-2.0
oneklc/dimod,oneklc/dimod
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. + """ + # we will override these in the __init__, but because they are abstract properties we need to + # signal that we are overriding them + edgelist = None + nodelist = None + children = None - todo - """ def __init__(self, sampler, nodelist, edgelist): - Sampler.__init__(self) - Composite.__init__(self, sampler) - Structured.__init__(self, nodelist, edgelist) + self.children = [sampler] + self.nodelist = nodelist + self.edgelist = edgelist + + @property + def parameters(self): + return self.child.parameters + + @property + def properties(self): + return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
Update Structure composite to use the new abc
## Code Before: from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs) ## Instruction: Update Structure composite to use the new abc ## Code After: from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
// ... existing code ... class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured // ... rest of the code ...
bb578d4237ccaf16fe5c38842cc100cdbefc0119
senlin/tests/functional/drivers/openstack/__init__.py
senlin/tests/functional/drivers/openstack/__init__.py
from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 from senlin.tests.functional.drivers.openstack import nova_v2 # Currently, only fake nova_v2 driver is supported def compute(params): return nova_v2.NovaClient(params) def loadbalancing(params): return lbaas.LoadBalancerDriver(params) def network(params): return neutron_v2.NeutronClient(params) def orchestration(params): return heat_v1.HeatClient(params) def telemetry(params): return ceilometer_v2.CeilometerClient(params)
from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 from senlin.drivers.openstack import keystone_v3 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 from senlin.tests.functional.drivers.openstack import nova_v2 # Currently, only fake nova_v2 driver is supported compute = nova_v2.NovaClient identity = keystone_v3.KeystoneClient loadbalancing = lbaas.LoadBalancerDriver network = neutron_v2.NeutronClient orchestration = heat_v1.HeatClient telemetry = ceilometer_v2.CeilometerClient
Add keystone driver plugin for functional test
Add keystone driver plugin for functional test This patch adds keystone driver plugin for functional test. Change-Id: Iefa9c1b8956854ae75f672627aa3d2f9f7d22c0e
Python
apache-2.0
openstack/senlin,stackforge/senlin,tengqm/senlin-container,stackforge/senlin,openstack/senlin,tengqm/senlin-container,Alzon/senlin,Alzon/senlin,openstack/senlin
from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 + from senlin.drivers.openstack import keystone_v3 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 from senlin.tests.functional.drivers.openstack import nova_v2 # Currently, only fake nova_v2 driver is supported - def compute(params): - return nova_v2.NovaClient(params) + compute = nova_v2.NovaClient + identity = keystone_v3.KeystoneClient + loadbalancing = lbaas.LoadBalancerDriver + network = neutron_v2.NeutronClient + orchestration = heat_v1.HeatClient + telemetry = ceilometer_v2.CeilometerClient - - def loadbalancing(params): - return lbaas.LoadBalancerDriver(params) - - - def network(params): - return neutron_v2.NeutronClient(params) - - - def orchestration(params): - return heat_v1.HeatClient(params) - - - def telemetry(params): - return ceilometer_v2.CeilometerClient(params) -
Add keystone driver plugin for functional test
## Code Before: from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 from senlin.tests.functional.drivers.openstack import nova_v2 # Currently, only fake nova_v2 driver is supported def compute(params): return nova_v2.NovaClient(params) def loadbalancing(params): return lbaas.LoadBalancerDriver(params) def network(params): return neutron_v2.NeutronClient(params) def orchestration(params): return heat_v1.HeatClient(params) def telemetry(params): return ceilometer_v2.CeilometerClient(params) ## Instruction: Add keystone driver plugin for functional test ## Code After: from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 from senlin.drivers.openstack import keystone_v3 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 from senlin.tests.functional.drivers.openstack import nova_v2 # Currently, only fake nova_v2 driver is supported compute = nova_v2.NovaClient identity = keystone_v3.KeystoneClient loadbalancing = lbaas.LoadBalancerDriver network = neutron_v2.NeutronClient orchestration = heat_v1.HeatClient telemetry = ceilometer_v2.CeilometerClient
// ... existing code ... from senlin.drivers.openstack import ceilometer_v2 from senlin.drivers.openstack import heat_v1 from senlin.drivers.openstack import keystone_v3 from senlin.drivers.openstack import lbaas from senlin.drivers.openstack import neutron_v2 // ... modified code ... # Currently, only fake nova_v2 driver is supported compute = nova_v2.NovaClient identity = keystone_v3.KeystoneClient loadbalancing = lbaas.LoadBalancerDriver network = neutron_v2.NeutronClient orchestration = heat_v1.HeatClient telemetry = ceilometer_v2.CeilometerClient // ... rest of the code ...
cc6ce477550152135eed5a9e35bca8144be10111
groupmestats/plotly_helpers.py
groupmestats/plotly_helpers.py
import plotly def try_saving_plotly_figure(figure, filename): try: plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: if 'The response from plotly could not be translated.'in str(e): print("Failed to save plotly figure. <home>/.plotly/.credentials" " might not be configured correctly? " "Or you may have hit your plotly account's rate limit" " (http://help.plot.ly/api-rate-limits/)") else: raise # A green bar with slightly darker green line marker = dict( color='#4BB541', line=dict( color='#3A9931', width=1.5, ) )
import plotly def try_saving_plotly_figure(figure, filename): try: print("Saving plot to '%s'" % filename) plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: if 'The response from plotly could not be translated.'in str(e): print("Failed to save plotly figure. <home>/.plotly/.credentials" " might not be configured correctly? " "Or you may have hit your plotly account's rate limit" " (http://help.plot.ly/api-rate-limits/)") else: raise # A green bar with slightly darker green line marker = dict( color='#4BB541', line=dict( color='#3A9931', width=1.5, ) )
Print when saving plot to file
Print when saving plot to file
Python
mit
kjteske/groupmestats,kjteske/groupmestats
import plotly def try_saving_plotly_figure(figure, filename): try: + print("Saving plot to '%s'" % filename) plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: if 'The response from plotly could not be translated.'in str(e): print("Failed to save plotly figure. <home>/.plotly/.credentials" " might not be configured correctly? " "Or you may have hit your plotly account's rate limit" " (http://help.plot.ly/api-rate-limits/)") else: raise # A green bar with slightly darker green line marker = dict( color='#4BB541', line=dict( color='#3A9931', width=1.5, ) )
Print when saving plot to file
## Code Before: import plotly def try_saving_plotly_figure(figure, filename): try: plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: if 'The response from plotly could not be translated.'in str(e): print("Failed to save plotly figure. <home>/.plotly/.credentials" " might not be configured correctly? " "Or you may have hit your plotly account's rate limit" " (http://help.plot.ly/api-rate-limits/)") else: raise # A green bar with slightly darker green line marker = dict( color='#4BB541', line=dict( color='#3A9931', width=1.5, ) ) ## Instruction: Print when saving plot to file ## Code After: import plotly def try_saving_plotly_figure(figure, filename): try: print("Saving plot to '%s'" % filename) plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: if 'The response from plotly could not be translated.'in str(e): print("Failed to save plotly figure. <home>/.plotly/.credentials" " might not be configured correctly? " "Or you may have hit your plotly account's rate limit" " (http://help.plot.ly/api-rate-limits/)") else: raise # A green bar with slightly darker green line marker = dict( color='#4BB541', line=dict( color='#3A9931', width=1.5, ) )
// ... existing code ... def try_saving_plotly_figure(figure, filename): try: print("Saving plot to '%s'" % filename) plotly.plotly.image.save_as(figure, filename) except plotly.exceptions.PlotlyError as e: // ... rest of the code ...
a5ff4c247030559c83a06976fcda062c0c42d810
django_fixmystreet/fixmystreet/tests/__init__.py
django_fixmystreet/fixmystreet/tests/__init__.py
import shutil import os from django.core.files.storage import default_storage from django.test import TestCase class SampleFilesTestCase(TestCase): fixtures = ['sample'] @classmethod def setUpClass(cls): default_storage.location = 'media' # force using source media folder to avoid real data erasing # @classmethod # def setUpClass(cls): # shutil.copytree('media', 'media-tmp') # default_storage.location = 'media-tmp' # # @classmethod # def tearDownClass(self): # shutil.rmtree('media-tmp') def _fixture_setup(self): if os.path.exists('media/photos'): shutil.rmtree('media/photos') shutil.copytree('media/photos-sample', 'media/photos') super(SampleFilesTestCase, self)._fixture_setup() def tearDown(self): shutil.rmtree('media/photos') from django_fixmystreet.fixmystreet.tests.views import * from django_fixmystreet.fixmystreet.tests.reports import * from django_fixmystreet.fixmystreet.tests.users import * from django_fixmystreet.fixmystreet.tests.organisation_entity import * from django_fixmystreet.fixmystreet.tests.mail import * # from django_fixmystreet.fixmystreet.tests.api import *
import shutil import os from django.core.files.storage import default_storage from django.test import TestCase class SampleFilesTestCase(TestCase): fixtures = ['sample'] @classmethod def setUpClass(cls): default_storage.location = 'media' # force using source media folder to avoid real data erasing # @classmethod # def setUpClass(cls): # shutil.copytree('media', 'media-tmp') # default_storage.location = 'media-tmp' # # @classmethod # def tearDownClass(self): # shutil.rmtree('media-tmp') def _fixture_setup(self): if os.path.exists('media/photos'): shutil.rmtree('media/photos') shutil.copytree('media/photos-sample', 'media/photos') super(SampleFilesTestCase, self)._fixture_setup() from django_fixmystreet.fixmystreet.tests.views import * from django_fixmystreet.fixmystreet.tests.reports import * from django_fixmystreet.fixmystreet.tests.users import * from django_fixmystreet.fixmystreet.tests.organisation_entity import * from django_fixmystreet.fixmystreet.tests.mail import * # from django_fixmystreet.fixmystreet.tests.api import *
Fix unit test fixtures files
Fix unit test fixtures files
Python
agpl-3.0
IMIO/django-fixmystreet,IMIO/django-fixmystreet,IMIO/django-fixmystreet,IMIO/django-fixmystreet
import shutil import os from django.core.files.storage import default_storage from django.test import TestCase class SampleFilesTestCase(TestCase): fixtures = ['sample'] @classmethod def setUpClass(cls): default_storage.location = 'media' # force using source media folder to avoid real data erasing # @classmethod # def setUpClass(cls): # shutil.copytree('media', 'media-tmp') # default_storage.location = 'media-tmp' # # @classmethod # def tearDownClass(self): # shutil.rmtree('media-tmp') def _fixture_setup(self): if os.path.exists('media/photos'): shutil.rmtree('media/photos') shutil.copytree('media/photos-sample', 'media/photos') super(SampleFilesTestCase, self)._fixture_setup() - def tearDown(self): - shutil.rmtree('media/photos') from django_fixmystreet.fixmystreet.tests.views import * from django_fixmystreet.fixmystreet.tests.reports import * from django_fixmystreet.fixmystreet.tests.users import * from django_fixmystreet.fixmystreet.tests.organisation_entity import * from django_fixmystreet.fixmystreet.tests.mail import * # from django_fixmystreet.fixmystreet.tests.api import *
Fix unit test fixtures files
## Code Before: import shutil import os from django.core.files.storage import default_storage from django.test import TestCase class SampleFilesTestCase(TestCase): fixtures = ['sample'] @classmethod def setUpClass(cls): default_storage.location = 'media' # force using source media folder to avoid real data erasing # @classmethod # def setUpClass(cls): # shutil.copytree('media', 'media-tmp') # default_storage.location = 'media-tmp' # # @classmethod # def tearDownClass(self): # shutil.rmtree('media-tmp') def _fixture_setup(self): if os.path.exists('media/photos'): shutil.rmtree('media/photos') shutil.copytree('media/photos-sample', 'media/photos') super(SampleFilesTestCase, self)._fixture_setup() def tearDown(self): shutil.rmtree('media/photos') from django_fixmystreet.fixmystreet.tests.views import * from django_fixmystreet.fixmystreet.tests.reports import * from django_fixmystreet.fixmystreet.tests.users import * from django_fixmystreet.fixmystreet.tests.organisation_entity import * from django_fixmystreet.fixmystreet.tests.mail import * # from django_fixmystreet.fixmystreet.tests.api import * ## Instruction: Fix unit test fixtures files ## Code After: import shutil import os from django.core.files.storage import default_storage from django.test import TestCase class SampleFilesTestCase(TestCase): fixtures = ['sample'] @classmethod def setUpClass(cls): default_storage.location = 'media' # force using source media folder to avoid real data erasing # @classmethod # def setUpClass(cls): # shutil.copytree('media', 'media-tmp') # default_storage.location = 'media-tmp' # # @classmethod # def tearDownClass(self): # shutil.rmtree('media-tmp') def _fixture_setup(self): if os.path.exists('media/photos'): shutil.rmtree('media/photos') shutil.copytree('media/photos-sample', 'media/photos') super(SampleFilesTestCase, self)._fixture_setup() from django_fixmystreet.fixmystreet.tests.views import * from django_fixmystreet.fixmystreet.tests.reports import * from django_fixmystreet.fixmystreet.tests.users import * from django_fixmystreet.fixmystreet.tests.organisation_entity import * from django_fixmystreet.fixmystreet.tests.mail import * # from django_fixmystreet.fixmystreet.tests.api import *
// ... existing code ... super(SampleFilesTestCase, self)._fixture_setup() from django_fixmystreet.fixmystreet.tests.views import * // ... rest of the code ...
bc8e548e51fddc251eb2e915883e3ee57bb9515b
zc_common/jwt_auth/utils.py
zc_common/jwt_auth/utils.py
import jwt from rest_framework_jwt.settings import api_settings def jwt_payload_handler(user): # The handler from rest_framework_jwt removed user_id, so this is a fork payload = { 'id': user.pk, 'roles': user.get_roles(), } return payload def jwt_encode_handler(payload): return jwt.encode( payload, api_settings.JWT_SECRET_KEY, api_settings.JWT_ALGORITHM ).decode('utf-8')
import jwt from rest_framework_jwt.settings import api_settings def jwt_payload_handler(user): '''Constructs a payload for a user JWT. This is a slimmed down version of https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11 :param User: an object with `pk` and `get_roles()` :return: A dictionary that can be passed into `jwt_encode_handler` ''' payload = { 'id': user.pk, 'roles': user.get_roles(), } return payload def jwt_encode_handler(payload): ''' Encodes a payload into a valid JWT. :param payload: a dictionary :return: an encoded JWT string ''' return jwt.encode( payload, api_settings.JWT_SECRET_KEY, api_settings.JWT_ALGORITHM ).decode('utf-8')
Add docstrings to jwt handlers
Add docstrings to jwt handlers
Python
mit
ZeroCater/zc_common,ZeroCater/zc_common
import jwt from rest_framework_jwt.settings import api_settings def jwt_payload_handler(user): - # The handler from rest_framework_jwt removed user_id, so this is a fork + '''Constructs a payload for a user JWT. This is a slimmed down version of + https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11 + + :param User: an object with `pk` and `get_roles()` + :return: A dictionary that can be passed into `jwt_encode_handler` + ''' + payload = { 'id': user.pk, 'roles': user.get_roles(), } return payload def jwt_encode_handler(payload): + ''' + Encodes a payload into a valid JWT. + + :param payload: a dictionary + :return: an encoded JWT string + ''' + return jwt.encode( payload, api_settings.JWT_SECRET_KEY, api_settings.JWT_ALGORITHM ).decode('utf-8')
Add docstrings to jwt handlers
## Code Before: import jwt from rest_framework_jwt.settings import api_settings def jwt_payload_handler(user): # The handler from rest_framework_jwt removed user_id, so this is a fork payload = { 'id': user.pk, 'roles': user.get_roles(), } return payload def jwt_encode_handler(payload): return jwt.encode( payload, api_settings.JWT_SECRET_KEY, api_settings.JWT_ALGORITHM ).decode('utf-8') ## Instruction: Add docstrings to jwt handlers ## Code After: import jwt from rest_framework_jwt.settings import api_settings def jwt_payload_handler(user): '''Constructs a payload for a user JWT. This is a slimmed down version of https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11 :param User: an object with `pk` and `get_roles()` :return: A dictionary that can be passed into `jwt_encode_handler` ''' payload = { 'id': user.pk, 'roles': user.get_roles(), } return payload def jwt_encode_handler(payload): ''' Encodes a payload into a valid JWT. :param payload: a dictionary :return: an encoded JWT string ''' return jwt.encode( payload, api_settings.JWT_SECRET_KEY, api_settings.JWT_ALGORITHM ).decode('utf-8')
# ... existing code ... def jwt_payload_handler(user): '''Constructs a payload for a user JWT. This is a slimmed down version of https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11 :param User: an object with `pk` and `get_roles()` :return: A dictionary that can be passed into `jwt_encode_handler` ''' payload = { 'id': user.pk, # ... modified code ... def jwt_encode_handler(payload): ''' Encodes a payload into a valid JWT. :param payload: a dictionary :return: an encoded JWT string ''' return jwt.encode( payload, # ... rest of the code ...
0eca2340a4d38b542659505ba386a23129f8ac0b
google/cloud/speech/__init__.py
google/cloud/speech/__init__.py
"""Google Cloud Speech API wrapper.""" from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection
"""Google Cloud Speech API wrapper.""" from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection from google.cloud.speech.encoding import Encoding
Make Encoding accessible from speech.Encoding.
Make Encoding accessible from speech.Encoding.
Python
apache-2.0
googleapis/python-speech,googleapis/python-speech
"""Google Cloud Speech API wrapper.""" from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection + from google.cloud.speech.encoding import Encoding
Make Encoding accessible from speech.Encoding.
## Code Before: """Google Cloud Speech API wrapper.""" from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection ## Instruction: Make Encoding accessible from speech.Encoding. ## Code After: """Google Cloud Speech API wrapper.""" from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection from google.cloud.speech.encoding import Encoding
// ... existing code ... from google.cloud.speech.client import Client from google.cloud.speech.connection import Connection from google.cloud.speech.encoding import Encoding // ... rest of the code ...
4b84cedd15a2774391544a6edee3532e5e267608
tests/docs/test_docs.py
tests/docs/test_docs.py
import subprocess import unittest import os import subprocess import unittest import os class Doc_Test(unittest.TestCase): @property def path_to_docs(self): dirname, filename = os.path.split(os.path.abspath(__file__)) return dirname.split(os.path.sep)[:-2] + ['docs'] def test_html(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) os.chdir(wd) def test_linkcheck(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "linkcheck"]) print(response.returncode) self.assertTrue(response.returncode == 0) os.chdir(wd) if __name__ == '__main__': unittest.main()
import subprocess import unittest import os import subprocess import unittest import os class Doc_Test(unittest.TestCase): @property def path_to_docs(self): dirname, filename = os.path.split(os.path.abspath(__file__)) return dirname.split(os.path.sep)[:-2] + ['docs'] def test_html(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows # self.assertTrue(response == 0) os.chdir(wd) def test_linkcheck(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "linkcheck"]) print(response.returncode) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows # print(response) # self.assertTrue(response == 0) os.chdir(wd) if __name__ == '__main__': unittest.main()
Edit docs test for local test on windows machine
Edit docs test for local test on windows machine
Python
mit
simpeg/simpeg
import subprocess import unittest import os import subprocess import unittest import os class Doc_Test(unittest.TestCase): @property def path_to_docs(self): dirname, filename = os.path.split(os.path.abspath(__file__)) return dirname.split(os.path.sep)[:-2] + ['docs'] def test_html(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) + # response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows + # self.assertTrue(response == 0) + os.chdir(wd) def test_linkcheck(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "linkcheck"]) print(response.returncode) self.assertTrue(response.returncode == 0) + # response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows + # print(response) + # self.assertTrue(response == 0) + os.chdir(wd) if __name__ == '__main__': unittest.main()
Edit docs test for local test on windows machine
## Code Before: import subprocess import unittest import os import subprocess import unittest import os class Doc_Test(unittest.TestCase): @property def path_to_docs(self): dirname, filename = os.path.split(os.path.abspath(__file__)) return dirname.split(os.path.sep)[:-2] + ['docs'] def test_html(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) os.chdir(wd) def test_linkcheck(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "linkcheck"]) print(response.returncode) self.assertTrue(response.returncode == 0) os.chdir(wd) if __name__ == '__main__': unittest.main() ## Instruction: Edit docs test for local test on windows machine ## Code After: import subprocess import unittest import os import subprocess import unittest import os class Doc_Test(unittest.TestCase): @property def path_to_docs(self): dirname, filename = os.path.split(os.path.abspath(__file__)) return dirname.split(os.path.sep)[:-2] + ['docs'] def test_html(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows # self.assertTrue(response == 0) os.chdir(wd) def test_linkcheck(self): wd = os.getcwd() os.chdir(os.path.sep.join(self.path_to_docs)) response = subprocess.run(["make", "linkcheck"]) print(response.returncode) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows # print(response) # self.assertTrue(response == 0) os.chdir(wd) if __name__ == '__main__': unittest.main()
... response = subprocess.run(["make", "html"]) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows # self.assertTrue(response == 0) os.chdir(wd) ... print(response.returncode) self.assertTrue(response.returncode == 0) # response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows # print(response) # self.assertTrue(response == 0) os.chdir(wd) ...
7d3de3aa2441739aa951aa100c057cfa878887d5
nukedb.py
nukedb.py
import sqlite3 if __name__=="__main__": conn = sqlite3.connect('auxgis.db') c = conn.cursor() try: c.execute('''DROP TABLE pos;''') except: pass try: c.execute('''DROP TABLE data;''') except: pass conn.commit()
import sqlite3 if __name__=="__main__": conn = sqlite3.connect('auxgis.db') c = conn.cursor() try: c.execute('''DROP TABLE pos;''') except: pass try: c.execute('''DROP TABLE data;''') except: pass try: c.execute('''DROP TABLE recentchanges;''') except: pass conn.commit()
Drop recent changes on nuke
Drop recent changes on nuke
Python
bsd-3-clause
TimSC/auxgis
import sqlite3 if __name__=="__main__": conn = sqlite3.connect('auxgis.db') c = conn.cursor() try: c.execute('''DROP TABLE pos;''') except: pass try: c.execute('''DROP TABLE data;''') except: pass + try: + c.execute('''DROP TABLE recentchanges;''') + except: + pass conn.commit()
Drop recent changes on nuke
## Code Before: import sqlite3 if __name__=="__main__": conn = sqlite3.connect('auxgis.db') c = conn.cursor() try: c.execute('''DROP TABLE pos;''') except: pass try: c.execute('''DROP TABLE data;''') except: pass conn.commit() ## Instruction: Drop recent changes on nuke ## Code After: import sqlite3 if __name__=="__main__": conn = sqlite3.connect('auxgis.db') c = conn.cursor() try: c.execute('''DROP TABLE pos;''') except: pass try: c.execute('''DROP TABLE data;''') except: pass try: c.execute('''DROP TABLE recentchanges;''') except: pass conn.commit()
// ... existing code ... except: pass try: c.execute('''DROP TABLE recentchanges;''') except: pass conn.commit() // ... rest of the code ...
4922d53f95b3f7c055afe1d0af0088b505cbc0d2
addons/bestja_configuration_ucw/__openerp__.py
addons/bestja_configuration_ucw/__openerp__.py
{ 'name': "Bestja: UCW", 'summary': "Installation configuration for UCW", 'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu", 'author': "Laboratorium EE", 'website': "http://www.laboratorium.ee", 'version': '0.1', 'category': 'Specific Industry Applications', 'depends': [ 'base', 'bestja_base', 'bestja_volunteer', 'bestja_volunteer_notes', 'bestja_account_deletion', 'bestja_organization', 'bestja_project', 'bestja_offers', 'bestja_offers_moderation', 'bestja_offers_invitations', 'bestja_offers_categorization', 'bestja_files', 'bestja_application_moderation', 'bestja_ucw_permissions', ], 'data': [ 'data.xml', ], 'application': True, }
{ 'name': "Bestja: UCW", 'summary': "Installation configuration for UCW", 'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu", 'author': "Laboratorium EE", 'website': "http://www.laboratorium.ee", 'version': '0.1', 'category': 'Specific Industry Applications', 'depends': [ 'base', 'website_blog', 'bestja_base', 'bestja_volunteer', 'bestja_volunteer_notes', 'bestja_account_deletion', 'bestja_organization', 'bestja_project', 'bestja_offers', 'bestja_offers_moderation', 'bestja_offers_invitations', 'bestja_offers_categorization', 'bestja_files', 'bestja_application_moderation', 'bestja_ucw_permissions', ], 'data': [ 'data.xml', ], 'application': True, }
Enable Odoo blog for UCW
Enable Odoo blog for UCW
Python
agpl-3.0
EE/bestja,EE/bestja,KamilWo/bestja,KamilWo/bestja,KrzysiekJ/bestja,ludwiktrammer/bestja,EE/bestja,ludwiktrammer/bestja,KamilWo/bestja,KrzysiekJ/bestja,ludwiktrammer/bestja,KrzysiekJ/bestja
{ 'name': "Bestja: UCW", 'summary': "Installation configuration for UCW", 'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu", 'author': "Laboratorium EE", 'website': "http://www.laboratorium.ee", 'version': '0.1', 'category': 'Specific Industry Applications', 'depends': [ 'base', + 'website_blog', 'bestja_base', 'bestja_volunteer', 'bestja_volunteer_notes', 'bestja_account_deletion', 'bestja_organization', 'bestja_project', 'bestja_offers', 'bestja_offers_moderation', 'bestja_offers_invitations', 'bestja_offers_categorization', 'bestja_files', 'bestja_application_moderation', 'bestja_ucw_permissions', ], 'data': [ 'data.xml', ], 'application': True, }
Enable Odoo blog for UCW
## Code Before: { 'name': "Bestja: UCW", 'summary': "Installation configuration for UCW", 'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu", 'author': "Laboratorium EE", 'website': "http://www.laboratorium.ee", 'version': '0.1', 'category': 'Specific Industry Applications', 'depends': [ 'base', 'bestja_base', 'bestja_volunteer', 'bestja_volunteer_notes', 'bestja_account_deletion', 'bestja_organization', 'bestja_project', 'bestja_offers', 'bestja_offers_moderation', 'bestja_offers_invitations', 'bestja_offers_categorization', 'bestja_files', 'bestja_application_moderation', 'bestja_ucw_permissions', ], 'data': [ 'data.xml', ], 'application': True, } ## Instruction: Enable Odoo blog for UCW ## Code After: { 'name': "Bestja: UCW", 'summary': "Installation configuration for UCW", 'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu", 'author': "Laboratorium EE", 'website': "http://www.laboratorium.ee", 'version': '0.1', 'category': 'Specific Industry Applications', 'depends': [ 'base', 'website_blog', 'bestja_base', 'bestja_volunteer', 'bestja_volunteer_notes', 'bestja_account_deletion', 'bestja_organization', 'bestja_project', 'bestja_offers', 'bestja_offers_moderation', 'bestja_offers_invitations', 'bestja_offers_categorization', 'bestja_files', 'bestja_application_moderation', 'bestja_ucw_permissions', ], 'data': [ 'data.xml', ], 'application': True, }
// ... existing code ... 'depends': [ 'base', 'website_blog', 'bestja_base', 'bestja_volunteer', // ... rest of the code ...
8a7837a8ce7b35c3141374c6a5c99361261fa70a
Cura/avr_isp/chipDB.py
Cura/avr_isp/chipDB.py
avrChipDB = { 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
avrChipDB = { 'ATMega1280': { 'signature': [0x1E, 0x97, 0x03], 'pageSize': 128, 'pageCount': 512, }, 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
Add ATMega1280 chip to programmer chips.
Add ATMega1280 chip to programmer chips.
Python
agpl-3.0
MolarAmbiguity/OctoPrint,EZ3-India/EZ-Remote,JackGavin13/octoprint-test-not-finished,spapadim/OctoPrint,dragondgold/OctoPrint,hudbrog/OctoPrint,CapnBry/OctoPrint,Javierma/OctoPrint-TFG,chriskoz/OctoPrint,javivi001/OctoPrint,shohei/Octoprint,eddieparker/OctoPrint,MolarAmbiguity/OctoPrint,mayoff/OctoPrint,uuv/OctoPrint,C-o-r-E/OctoPrint,Mikk36/OctoPrint,DanLipsitt/OctoPrint,shohei/Octoprint,beeverycreative/BEEweb,alex1818/OctoPrint,EZ3-India/EZ-Remote,alex1818/OctoPrint,shohei/Octoprint,markwal/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,nicanor-romero/OctoPrint,punkkeks/OctoPrint,d42/octoprint-fork,Javierma/OctoPrint-TFG,3dprintcanalhouse/octoprint2,ErikDeBruijn/OctoPrint,punkkeks/OctoPrint,masterhou/OctoPrint,shaggythesheep/OctoPrint,chriskoz/OctoPrint,madhuni/AstroBox,Catrodigious/OctoPrint-TAM,alephobjects/Cura,javivi001/OctoPrint,uuv/OctoPrint,leductan-nguyen/RaionPi,MoonshineSG/OctoPrint,eliasbakken/OctoPrint,nicanor-romero/OctoPrint,Skeen/OctoPrint,javivi001/OctoPrint,Salandora/OctoPrint,jneves/OctoPrint,hudbrog/OctoPrint,shaggythesheep/OctoPrint,MoonshineSG/OctoPrint,skieast/OctoPrint,abinashk-inf/AstroBox,nickverschoor/OctoPrint,eddieparker/OctoPrint,EZ3-India/EZ-Remote,EZ3-India/EZ-Remote,abinashk-inf/AstroBox,mrbeam/OctoPrint,abinashk-inf/AstroBox,mrbeam/OctoPrint,Voxel8/OctoPrint,sstocker46/OctoPrint,bicephale/OctoPrint,dragondgold/OctoPrint,Jaesin/OctoPrint,mcanes/OctoPrint,ryanneufeld/OctoPrint,Salandora/OctoPrint,CapnBry/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,alephobjects/Cura,mcanes/OctoPrint,markwal/OctoPrint,sstocker46/OctoPrint,Jaesin/OctoPrint,3dprintcanalhouse/octoprint1,skieast/OctoPrint,madhuni/AstroBox,markwal/OctoPrint,Mikk36/OctoPrint,AstroPrint/AstroBox,ymilord/OctoPrint-MrBeam,dansantee/OctoPrint,Jaesin/OctoPrint,punkkeks/OctoPrint,ymilord/OctoPrint-MrBeam,rurkowce/octoprint-fork,foosel/OctoPrint,Salandora/OctoPrint,spapadim/OctoPrint,MoonshineSG/OctoPrint,spapadim/OctoPrint,madhuni/AstroBox,masterhou/OctoPrint,ymilord/OctoPrint-MrBeam,alephobjects/Cura,ryanneufeld/OctoPrint,chriskoz/OctoPrint,hudbrog/OctoPrint,Mikk36/OctoPrint,eddieparker/OctoPrint,leductan-nguyen/RaionPi,JackGavin13/octoprint-test-not-finished,beeverycreative/BEEweb,bicephale/OctoPrint,nicanor-romero/OctoPrint,jneves/OctoPrint,JackGavin13/octoprint-test-not-finished,ErikDeBruijn/OctoPrint,leductan-nguyen/RaionPi,CapnBry/OctoPrint,chriskoz/OctoPrint,ryanneufeld/OctoPrint,3dprintcanalhouse/octoprint1,mrbeam/OctoPrint,senttech/OctoPrint,Javierma/OctoPrint-TFG,dansantee/OctoPrint,Voxel8/OctoPrint,bicephale/OctoPrint,MolarAmbiguity/OctoPrint,MaxOLydian/OctoPrint,eliasbakken/OctoPrint,DanLipsitt/OctoPrint,mayoff/OctoPrint,Skeen/OctoPrint,Jaesin/OctoPrint,rurkowce/octoprint-fork,CapnBry/OctoPrint,AstroPrint/AstroBox,madhuni/AstroBox,uuv/OctoPrint,abinashk-inf/AstroBox,JackGavin13/octoprint-test-not-finished,SeveQ/OctoPrint,sstocker46/OctoPrint,dansantee/OctoPrint,skieast/OctoPrint,mayoff/OctoPrint,C-o-r-E/OctoPrint,eliasbakken/OctoPrint,ryanneufeld/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,bicephale/OctoPrint,SeveQ/OctoPrint,MoonshineSG/OctoPrint,SeveQ/OctoPrint,senttech/OctoPrint,shohei/Octoprint,ymilord/OctoPrint-MrBeam,3dprintcanalhouse/octoprint2,d42/octoprint-fork,mcanes/OctoPrint,Voxel8/OctoPrint,senttech/OctoPrint,ymilord/OctoPrint-MrBeam,leductan-nguyen/RaionPi,Javierma/OctoPrint-TFG,Salandora/OctoPrint,C-o-r-E/OctoPrint,alex1818/OctoPrint,MaxOLydian/OctoPrint,shaggythesheep/OctoPrint,masterhou/OctoPrint,shohei/Octoprint,ErikDeBruijn/OctoPrint,jneves/OctoPrint,Catrodigious/OctoPrint-TAM,foosel/OctoPrint,dragondgold/OctoPrint,senttech/OctoPrint,aerickson/OctoPrint,MaxOLydian/OctoPrint,nickverschoor/OctoPrint,Skeen/OctoPrint,Catrodigious/OctoPrint-TAM,AstroPrint/AstroBox
avrChipDB = { + 'ATMega1280': { + 'signature': [0x1E, 0x97, 0x03], + 'pageSize': 128, + 'pageCount': 512, + }, 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
Add ATMega1280 chip to programmer chips.
## Code Before: avrChipDB = { 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False ## Instruction: Add ATMega1280 chip to programmer chips. ## Code After: avrChipDB = { 'ATMega1280': { 'signature': [0x1E, 0x97, 0x03], 'pageSize': 128, 'pageCount': 512, }, 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
// ... existing code ... avrChipDB = { 'ATMega1280': { 'signature': [0x1E, 0x97, 0x03], 'pageSize': 128, 'pageCount': 512, }, 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], // ... rest of the code ...
874a6eff186d1c1ca6f90d69fd24fad11180c5a9
thread_output_ctrl.py
thread_output_ctrl.py
import threading import wx from styled_text_ctrl import StyledTextCtrl class ThreadOutputCtrl(StyledTextCtrl): def __init__(self, parent, env, auto_scroll=False): StyledTextCtrl.__init__(self, parent, env) self.auto_scroll = auto_scroll self.__lock = threading.Lock() self.__queue = [] self.__timer = wx.Timer(self) self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer) def __OnTimer(self, evt): self.flush() def flush(self): with self.__lock: queue, self.__queue = self.__queue, [] lines = "".join(queue) if lines: with self.ModifyReadOnly(): self.AppendText(lines) self.EmptyUndoBuffer() if self.auto_scroll: self.ScrollToLine(self.GetLineCount() - 1) def start(self, interval=100): self.SetReadOnly(True) self.__timer.Start(interval) def stop(self): self.__timer.Stop() self.flush() self.SetReadOnly(False) def write(self, s): with self.__lock: self.__queue.append(s) def ClearAll(self): with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self)
import threading import wx from styled_text_ctrl import StyledTextCtrl class ThreadOutputCtrl(StyledTextCtrl): def __init__(self, parent, env, auto_scroll=False): StyledTextCtrl.__init__(self, parent, env) self.auto_scroll = auto_scroll self.__lock = threading.Lock() self.__queue = [] self.__timer = wx.Timer(self) self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer) def __OnTimer(self, evt): self.flush() def flush(self): with self.__lock: queue, self.__queue = self.__queue, [] lines = "".join(queue) if lines: with self.ModifyReadOnly(): self.AppendText(lines) self.EmptyUndoBuffer() if self.auto_scroll: self.ScrollToLine(self.GetLineCount() - 1) def start(self, interval=100): self.SetReadOnly(True) self.__timer.Start(interval) def stop(self): self.__timer.Stop() self.flush() self.SetReadOnly(False) def write(self, s): with self.__lock: self.__queue.append(s) def ClearAll(self): with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self) self.EmptyUndoBuffer()
Clear undo buffer when terminal cleared.
Clear undo buffer when terminal cleared.
Python
mit
shaurz/devo
import threading import wx from styled_text_ctrl import StyledTextCtrl class ThreadOutputCtrl(StyledTextCtrl): def __init__(self, parent, env, auto_scroll=False): StyledTextCtrl.__init__(self, parent, env) self.auto_scroll = auto_scroll self.__lock = threading.Lock() self.__queue = [] self.__timer = wx.Timer(self) self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer) def __OnTimer(self, evt): self.flush() def flush(self): with self.__lock: queue, self.__queue = self.__queue, [] lines = "".join(queue) if lines: with self.ModifyReadOnly(): self.AppendText(lines) self.EmptyUndoBuffer() if self.auto_scroll: self.ScrollToLine(self.GetLineCount() - 1) def start(self, interval=100): self.SetReadOnly(True) self.__timer.Start(interval) def stop(self): self.__timer.Stop() self.flush() self.SetReadOnly(False) def write(self, s): with self.__lock: self.__queue.append(s) def ClearAll(self): with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self) + self.EmptyUndoBuffer()
Clear undo buffer when terminal cleared.
## Code Before: import threading import wx from styled_text_ctrl import StyledTextCtrl class ThreadOutputCtrl(StyledTextCtrl): def __init__(self, parent, env, auto_scroll=False): StyledTextCtrl.__init__(self, parent, env) self.auto_scroll = auto_scroll self.__lock = threading.Lock() self.__queue = [] self.__timer = wx.Timer(self) self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer) def __OnTimer(self, evt): self.flush() def flush(self): with self.__lock: queue, self.__queue = self.__queue, [] lines = "".join(queue) if lines: with self.ModifyReadOnly(): self.AppendText(lines) self.EmptyUndoBuffer() if self.auto_scroll: self.ScrollToLine(self.GetLineCount() - 1) def start(self, interval=100): self.SetReadOnly(True) self.__timer.Start(interval) def stop(self): self.__timer.Stop() self.flush() self.SetReadOnly(False) def write(self, s): with self.__lock: self.__queue.append(s) def ClearAll(self): with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self) ## Instruction: Clear undo buffer when terminal cleared. ## Code After: import threading import wx from styled_text_ctrl import StyledTextCtrl class ThreadOutputCtrl(StyledTextCtrl): def __init__(self, parent, env, auto_scroll=False): StyledTextCtrl.__init__(self, parent, env) self.auto_scroll = auto_scroll self.__lock = threading.Lock() self.__queue = [] self.__timer = wx.Timer(self) self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer) def __OnTimer(self, evt): self.flush() def flush(self): with self.__lock: queue, self.__queue = self.__queue, [] lines = "".join(queue) if lines: with self.ModifyReadOnly(): self.AppendText(lines) self.EmptyUndoBuffer() if self.auto_scroll: self.ScrollToLine(self.GetLineCount() - 1) def start(self, interval=100): self.SetReadOnly(True) self.__timer.Start(interval) def stop(self): self.__timer.Stop() self.flush() self.SetReadOnly(False) def write(self, s): with self.__lock: self.__queue.append(s) def ClearAll(self): with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self) self.EmptyUndoBuffer()
... with self.ModifyReadOnly(): StyledTextCtrl.ClearAll(self) self.EmptyUndoBuffer() ...
4d7c1fec37943558ccc8bf6a17860b2a86fe1941
gee_asset_manager/batch_copy.py
gee_asset_manager/batch_copy.py
import ee import os import csv import logging def copy(source, destination): with open(source, 'r') as f: reader = csv.reader(f) for line in reader: name = line[0] gme_id = line[1] gme_path = 'GME/images/' + gme_id ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) ee.data.copyAsset(gme_path, ee_path) if __name__ == '__main__': ee.Initialize() assets = '/home/tracek/Data/consbio2016/test.csv' with open(assets, 'r') as f: reader = csv.reader(f)
import ee import os import csv import logging def copy(source, destination): with open(source, 'r') as f: reader = csv.reader(f) for line in reader: name = line[0] gme_id = line[1] gme_path = 'GME/images/' + gme_id ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) try: ee.data.copyAsset(gme_path, ee_path) except ee.EEException as e: with open('failed_batch_copy.csv', 'w') as fout: fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e) if __name__ == '__main__': ee.Initialize() assets = '/home/tracek/Data/consbio2016/test.csv' with open(assets, 'r') as f: reader = csv.reader(f)
Add exception handling to batch copy
Add exception handling to batch copy
Python
apache-2.0
tracek/gee_asset_manager
import ee import os import csv import logging def copy(source, destination): with open(source, 'r') as f: reader = csv.reader(f) for line in reader: name = line[0] gme_id = line[1] gme_path = 'GME/images/' + gme_id ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) + try: - ee.data.copyAsset(gme_path, ee_path) + ee.data.copyAsset(gme_path, ee_path) + except ee.EEException as e: + with open('failed_batch_copy.csv', 'w') as fout: + fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e) if __name__ == '__main__': ee.Initialize() assets = '/home/tracek/Data/consbio2016/test.csv' with open(assets, 'r') as f: reader = csv.reader(f)
Add exception handling to batch copy
## Code Before: import ee import os import csv import logging def copy(source, destination): with open(source, 'r') as f: reader = csv.reader(f) for line in reader: name = line[0] gme_id = line[1] gme_path = 'GME/images/' + gme_id ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) ee.data.copyAsset(gme_path, ee_path) if __name__ == '__main__': ee.Initialize() assets = '/home/tracek/Data/consbio2016/test.csv' with open(assets, 'r') as f: reader = csv.reader(f) ## Instruction: Add exception handling to batch copy ## Code After: import ee import os import csv import logging def copy(source, destination): with open(source, 'r') as f: reader = csv.reader(f) for line in reader: name = line[0] gme_id = line[1] gme_path = 'GME/images/' + gme_id ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) try: ee.data.copyAsset(gme_path, ee_path) except ee.EEException as e: with open('failed_batch_copy.csv', 'w') as fout: fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e) if __name__ == '__main__': ee.Initialize() assets = '/home/tracek/Data/consbio2016/test.csv' with open(assets, 'r') as f: reader = csv.reader(f)
# ... existing code ... ee_path = os.path.join(destination, name) logging.info('Copying asset %s to %s', gme_path, ee_path) try: ee.data.copyAsset(gme_path, ee_path) except ee.EEException as e: with open('failed_batch_copy.csv', 'w') as fout: fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e) # ... rest of the code ...
9b54d728a245855cba724a91d372a15a4f4abb6d
shop/checkout/models.py
shop/checkout/models.py
"""Checkout Models""" import functools from flask import redirect, url_for from fulfil_client.model import ModelType, StringType from shop.fulfilio import Model from shop.globals import current_cart, current_channel def not_empty_cart(function): @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.is_empty: return redirect(url_for('cart.view_cart')) return function(*args, **kwargs) return wrapper def sale_has_non_guest_party(function): """ Ensure that the sale has a party who is not guest. The sign-in method authomatically changes the party to a party based on the session. """ @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.sale and cart.sale.party and \ cart.sale.party.id == current_channel.anonymous_customer.id: return redirect(url_for('checkout.sign_in')) return function(*args, **kwargs) return wrapper class PaymentGateway(Model): __model_name__ = 'payment_gateway.gateway' provider = StringType() stripe_publishable_key = StringType() class PaymentProfile(Model): __model_name__ = 'party.payment_profile' party = ModelType('party.party') gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() rec_name = StringType()
"""Checkout Models""" import functools from flask import redirect, url_for from fulfil_client.model import ModelType, StringType from shop.fulfilio import Model from shop.globals import current_cart, current_channel def not_empty_cart(function): @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.is_empty: return redirect(url_for('cart.view_cart')) return function(*args, **kwargs) return wrapper def sale_has_non_guest_party(function): """ Ensure that the sale has a party who is not guest. The sign-in method authomatically changes the party to a party based on the session. """ @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.sale and cart.sale.party and \ cart.sale.party.id == current_channel.anonymous_customer.id: return redirect(url_for('checkout.sign_in')) return function(*args, **kwargs) return wrapper class PaymentGateway(Model): __model_name__ = 'payment_gateway.gateway' provider = StringType() stripe_publishable_key = StringType() class PaymentProfile(Model): __model_name__ = 'party.payment_profile' party = ModelType('party.party') gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() expiry_month = StringType() expiry_year = StringType() rec_name = StringType()
Add expiry fields on card model
Add expiry fields on card model
Python
bsd-3-clause
joeirimpan/shop,joeirimpan/shop,joeirimpan/shop
"""Checkout Models""" import functools from flask import redirect, url_for from fulfil_client.model import ModelType, StringType from shop.fulfilio import Model from shop.globals import current_cart, current_channel def not_empty_cart(function): @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.is_empty: return redirect(url_for('cart.view_cart')) return function(*args, **kwargs) return wrapper def sale_has_non_guest_party(function): """ Ensure that the sale has a party who is not guest. The sign-in method authomatically changes the party to a party based on the session. """ @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.sale and cart.sale.party and \ cart.sale.party.id == current_channel.anonymous_customer.id: return redirect(url_for('checkout.sign_in')) return function(*args, **kwargs) return wrapper class PaymentGateway(Model): __model_name__ = 'payment_gateway.gateway' provider = StringType() stripe_publishable_key = StringType() class PaymentProfile(Model): __model_name__ = 'party.payment_profile' party = ModelType('party.party') gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() + expiry_month = StringType() + expiry_year = StringType() rec_name = StringType()
Add expiry fields on card model
## Code Before: """Checkout Models""" import functools from flask import redirect, url_for from fulfil_client.model import ModelType, StringType from shop.fulfilio import Model from shop.globals import current_cart, current_channel def not_empty_cart(function): @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.is_empty: return redirect(url_for('cart.view_cart')) return function(*args, **kwargs) return wrapper def sale_has_non_guest_party(function): """ Ensure that the sale has a party who is not guest. The sign-in method authomatically changes the party to a party based on the session. """ @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.sale and cart.sale.party and \ cart.sale.party.id == current_channel.anonymous_customer.id: return redirect(url_for('checkout.sign_in')) return function(*args, **kwargs) return wrapper class PaymentGateway(Model): __model_name__ = 'payment_gateway.gateway' provider = StringType() stripe_publishable_key = StringType() class PaymentProfile(Model): __model_name__ = 'party.payment_profile' party = ModelType('party.party') gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() rec_name = StringType() ## Instruction: Add expiry fields on card model ## Code After: """Checkout Models""" import functools from flask import redirect, url_for from fulfil_client.model import ModelType, StringType from shop.fulfilio import Model from shop.globals import current_cart, current_channel def not_empty_cart(function): @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.is_empty: return redirect(url_for('cart.view_cart')) return function(*args, **kwargs) return wrapper def sale_has_non_guest_party(function): """ Ensure that the sale has a party who is not guest. The sign-in method authomatically changes the party to a party based on the session. """ @functools.wraps(function) def wrapper(*args, **kwargs): cart = current_cart if cart.sale and cart.sale.party and \ cart.sale.party.id == current_channel.anonymous_customer.id: return redirect(url_for('checkout.sign_in')) return function(*args, **kwargs) return wrapper class PaymentGateway(Model): __model_name__ = 'payment_gateway.gateway' provider = StringType() stripe_publishable_key = StringType() class PaymentProfile(Model): __model_name__ = 'party.payment_profile' party = ModelType('party.party') gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() expiry_month = StringType() expiry_year = StringType() rec_name = StringType()
// ... existing code ... gateway = ModelType('payment_gateway.gateway') last_4_digits = StringType() expiry_month = StringType() expiry_year = StringType() rec_name = StringType() // ... rest of the code ...
8df3076b6315a74e57ee27fe3478d36737be0ff9
roche/scripts/xml-load.py
roche/scripts/xml-load.py
import sys import os sys.path.append('.') import roche.settings from os import walk from eulexistdb.db import ExistDB from roche.settings import EXISTDB_SERVER_URL # # Timeout higher? # xmldb = ExistDB(timeout=30) xmldb.createCollection('docker', True) xmldb.createCollection('docker/texts', True) os.chdir('../dublin-store') for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'): xmldb.createCollection('docker/texts' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
import sys import os sys.path.append('.') import roche.settings from os import walk from eulexistdb.db import ExistDB from roche.settings import EXISTDB_SERVER_URL # # Timeout higher? # xmldb = ExistDB(timeout=30) xmldb.createCollection('docker', True) xmldb.createCollection('docker/texts', True) os.chdir('../dublin-store') for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'): xmldb.createCollection('docker/texts' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True) # # Load resources # for (dirpath, dirnames, filenames) in walk('resources'): xmldb.createCollection('docker' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
Load other resources into exist-db
Load other resources into exist-db
Python
mit
beijingren/roche-website,beijingren/roche-website,beijingren/roche-website,beijingren/roche-website
import sys import os sys.path.append('.') import roche.settings from os import walk from eulexistdb.db import ExistDB from roche.settings import EXISTDB_SERVER_URL # # Timeout higher? # xmldb = ExistDB(timeout=30) xmldb.createCollection('docker', True) xmldb.createCollection('docker/texts', True) os.chdir('../dublin-store') for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'): xmldb.createCollection('docker/texts' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True) + # + # Load resources + # + for (dirpath, dirnames, filenames) in walk('resources'): + xmldb.createCollection('docker' + '/' + dirpath, True) + if filenames: + for filename in filenames: + with open(dirpath + '/' + filename) as f: + xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True) +
Load other resources into exist-db
## Code Before: import sys import os sys.path.append('.') import roche.settings from os import walk from eulexistdb.db import ExistDB from roche.settings import EXISTDB_SERVER_URL # # Timeout higher? # xmldb = ExistDB(timeout=30) xmldb.createCollection('docker', True) xmldb.createCollection('docker/texts', True) os.chdir('../dublin-store') for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'): xmldb.createCollection('docker/texts' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True) ## Instruction: Load other resources into exist-db ## Code After: import sys import os sys.path.append('.') import roche.settings from os import walk from eulexistdb.db import ExistDB from roche.settings import EXISTDB_SERVER_URL # # Timeout higher? # xmldb = ExistDB(timeout=30) xmldb.createCollection('docker', True) xmldb.createCollection('docker/texts', True) os.chdir('../dublin-store') for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'): xmldb.createCollection('docker/texts' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True) # # Load resources # for (dirpath, dirnames, filenames) in walk('resources'): xmldb.createCollection('docker' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
# ... existing code ... with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True) # # Load resources # for (dirpath, dirnames, filenames) in walk('resources'): xmldb.createCollection('docker' + '/' + dirpath, True) if filenames: for filename in filenames: with open(dirpath + '/' + filename) as f: xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True) # ... rest of the code ...
61ca14440f39106b6109b96919b520e40170b1f3
examples/tour_examples/xkcd_tour.py
examples/tour_examples/xkcd_tour.py
from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_basic(self): self.open('https://xkcd.com/1117/') self.assert_element('img[alt="My Sky"]') self.create_shepherd_tour() self.add_tour_step("Welcome to XKCD!") self.add_tour_step("This is the XKCD logo.", "#masthead img") self.add_tour_step("Here's the daily webcomic.", "#comic img") self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') self.add_tour_step("Click for the license here.", 'a[rel="license"]') self.add_tour_step("This selects a random comic.", 'a[href*="random"]') self.add_tour_step("Thanks for taking this tour!") # self.export_tour() # Use this to export the tour as [my_tour.js] self.export_tour(filename="xkcd_tour.js") # You can customize the name self.play_tour()
from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_basic(self): self.open('https://xkcd.com/1117/') self.assert_element('img[alt="My Sky"]') self.create_shepherd_tour() self.add_tour_step("Welcome to XKCD!") self.add_tour_step("This is the XKCD logo.", "#masthead img") self.add_tour_step("Here's the daily webcomic.", "#comic img") self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') self.add_tour_step("Click here for the license.", 'a[rel="license"]') self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]') self.add_tour_step("Thanks for taking this tour!") self.export_tour(filename="xkcd_tour.js") # This exports the tour self.play_tour() # This plays the tour
Update a SeleniumBase tour example
Update a SeleniumBase tour example
Python
mit
mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase
from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_basic(self): self.open('https://xkcd.com/1117/') self.assert_element('img[alt="My Sky"]') self.create_shepherd_tour() self.add_tour_step("Welcome to XKCD!") self.add_tour_step("This is the XKCD logo.", "#masthead img") self.add_tour_step("Here's the daily webcomic.", "#comic img") self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') - self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]') + self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') - self.add_tour_step("Click for the license here.", 'a[rel="license"]') + self.add_tour_step("Click here for the license.", 'a[rel="license"]') - self.add_tour_step("This selects a random comic.", 'a[href*="random"]') + self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]') self.add_tour_step("Thanks for taking this tour!") - # self.export_tour() # Use this to export the tour as [my_tour.js] - self.export_tour(filename="xkcd_tour.js") # You can customize the name + self.export_tour(filename="xkcd_tour.js") # This exports the tour - self.play_tour() + self.play_tour() # This plays the tour
Update a SeleniumBase tour example
## Code Before: from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_basic(self): self.open('https://xkcd.com/1117/') self.assert_element('img[alt="My Sky"]') self.create_shepherd_tour() self.add_tour_step("Welcome to XKCD!") self.add_tour_step("This is the XKCD logo.", "#masthead img") self.add_tour_step("Here's the daily webcomic.", "#comic img") self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') self.add_tour_step("Click for the license here.", 'a[rel="license"]') self.add_tour_step("This selects a random comic.", 'a[href*="random"]') self.add_tour_step("Thanks for taking this tour!") # self.export_tour() # Use this to export the tour as [my_tour.js] self.export_tour(filename="xkcd_tour.js") # You can customize the name self.play_tour() ## Instruction: Update a SeleniumBase tour example ## Code After: from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_basic(self): self.open('https://xkcd.com/1117/') self.assert_element('img[alt="My Sky"]') self.create_shepherd_tour() self.add_tour_step("Welcome to XKCD!") self.add_tour_step("This is the XKCD logo.", "#masthead img") self.add_tour_step("Here's the daily webcomic.", "#comic img") self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') self.add_tour_step("Click here for the license.", 'a[rel="license"]') self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]') self.add_tour_step("Thanks for taking this tour!") self.export_tour(filename="xkcd_tour.js") # This exports the tour self.play_tour() # This plays the tour
# ... existing code ... self.add_tour_step("This is the title.", "#ctitle", alignment="top") self.add_tour_step("Click here for the next comic.", 'a[rel="next"]') self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]') self.add_tour_step("Learn about the author here.", 'a[rel="author"]') self.add_tour_step("Click here for the license.", 'a[rel="license"]') self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]') self.add_tour_step("Thanks for taking this tour!") self.export_tour(filename="xkcd_tour.js") # This exports the tour self.play_tour() # This plays the tour # ... rest of the code ...
5841f314636ee534342aa3e4530cc3ee933a052b
src/ezweb/compressor_filters.py
src/ezweb/compressor_filters.py
from compressor.filters import FilterBase class JSUseStrictFilter(FilterBase): def output(self, **kwargs): return self.remove_use_strict(self.content) def remove_use_strict(js): js = js.replace("'use strict';", '') js = js.replace('"use strict";', '') return js
from compressor.filters import FilterBase class JSUseStrictFilter(FilterBase): def output(self, **kwargs): return self.remove_use_strict(self.content) def remove_use_strict(self, js): # Replacing by a ';' is safer than replacing by '' js = js.replace("'use strict';", ';') js = js.replace('"use strict";', ';') return js
Fix a bug while replacing "use strict" JS pragmas
Fix a bug while replacing "use strict" JS pragmas
Python
agpl-3.0
jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,rockneurotiko/wirecloud
from compressor.filters import FilterBase class JSUseStrictFilter(FilterBase): def output(self, **kwargs): return self.remove_use_strict(self.content) - def remove_use_strict(js): + def remove_use_strict(self, js): + # Replacing by a ';' is safer than replacing by '' - js = js.replace("'use strict';", '') + js = js.replace("'use strict';", ';') - js = js.replace('"use strict";', '') + js = js.replace('"use strict";', ';') return js
Fix a bug while replacing "use strict" JS pragmas
## Code Before: from compressor.filters import FilterBase class JSUseStrictFilter(FilterBase): def output(self, **kwargs): return self.remove_use_strict(self.content) def remove_use_strict(js): js = js.replace("'use strict';", '') js = js.replace('"use strict";', '') return js ## Instruction: Fix a bug while replacing "use strict" JS pragmas ## Code After: from compressor.filters import FilterBase class JSUseStrictFilter(FilterBase): def output(self, **kwargs): return self.remove_use_strict(self.content) def remove_use_strict(self, js): # Replacing by a ';' is safer than replacing by '' js = js.replace("'use strict';", ';') js = js.replace('"use strict";', ';') return js
// ... existing code ... return self.remove_use_strict(self.content) def remove_use_strict(self, js): # Replacing by a ';' is safer than replacing by '' js = js.replace("'use strict';", ';') js = js.replace('"use strict";', ';') return js // ... rest of the code ...
7a552161eab19d24b7b221635e51a915adff0166
templater.py
templater.py
import string if __name__ == "__main__": import sys template_file = sys.argv[1] with open(template_file) as f: data = f.read() template = string.Template(data) template_mapping = {} for item in sys.argv[2:]: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) key, value = item.split("=", 1) template_mapping[key] = value print("-> Using mapping: %s" % str(template_mapping)) result = template.substitute(template_mapping) print("-----\n") print(result)
import string import os if __name__ == "__main__": from optparse import OptionParser parser = OptionParser() parser.add_option("-t", "--template", dest="template_file", help="Input template file") (options, args) = parser.parse_args() if not os.path.isfile(options.template_file): sys.stderr.write("Invalid input template file") exit(1) with open(options.template_file) as f: data = f.read() template = string.Template(data) template_mapping = {} for item in args: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) key, value = item.split("=", 1) template_mapping[key] = value print("-> Using mapping: %s" % str(template_mapping)) result = template.substitute(template_mapping) print("-----\n") print(result)
Use OptionParser instead of simple sys.argv.
Use OptionParser instead of simple sys.argv.
Python
mit
elecro/strep
import string + import os + if __name__ == "__main__": - import sys + from optparse import OptionParser - template_file = sys.argv[1] + parser = OptionParser() + parser.add_option("-t", "--template", dest="template_file", + help="Input template file") + (options, args) = parser.parse_args() + + if not os.path.isfile(options.template_file): + sys.stderr.write("Invalid input template file") + exit(1) + - with open(template_file) as f: + with open(options.template_file) as f: data = f.read() template = string.Template(data) template_mapping = {} - for item in sys.argv[2:]: + for item in args: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) key, value = item.split("=", 1) template_mapping[key] = value print("-> Using mapping: %s" % str(template_mapping)) result = template.substitute(template_mapping) print("-----\n") print(result)
Use OptionParser instead of simple sys.argv.
## Code Before: import string if __name__ == "__main__": import sys template_file = sys.argv[1] with open(template_file) as f: data = f.read() template = string.Template(data) template_mapping = {} for item in sys.argv[2:]: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) key, value = item.split("=", 1) template_mapping[key] = value print("-> Using mapping: %s" % str(template_mapping)) result = template.substitute(template_mapping) print("-----\n") print(result) ## Instruction: Use OptionParser instead of simple sys.argv. ## Code After: import string import os if __name__ == "__main__": from optparse import OptionParser parser = OptionParser() parser.add_option("-t", "--template", dest="template_file", help="Input template file") (options, args) = parser.parse_args() if not os.path.isfile(options.template_file): sys.stderr.write("Invalid input template file") exit(1) with open(options.template_file) as f: data = f.read() template = string.Template(data) template_mapping = {} for item in args: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) key, value = item.split("=", 1) template_mapping[key] = value print("-> Using mapping: %s" % str(template_mapping)) result = template.substitute(template_mapping) print("-----\n") print(result)
... import string import os if __name__ == "__main__": from optparse import OptionParser parser = OptionParser() parser.add_option("-t", "--template", dest="template_file", help="Input template file") (options, args) = parser.parse_args() if not os.path.isfile(options.template_file): sys.stderr.write("Invalid input template file") exit(1) with open(options.template_file) as f: data = f.read() ... template_mapping = {} for item in args: # item is in the following form: KEY=VALUE print("-> Current replacer %s" % item) ...
6decf1f48e56832b1d15d3fc26d92f9813d13353
coop_cms/moves.py
coop_cms/moves.py
import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
Fix HTMLParser compatibility in Python 3
Fix HTMLParser compatibility in Python 3
Python
bsd-3-clause
ljean/coop_cms,ljean/coop_cms,ljean/coop_cms
import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 + + from StringIO import StringIO + from HTMLParser import HTMLParser - from StringIO import StringIO + + + else: # Python 3 - from html.parser import HTMLParser from io import BytesIO as StringIO + + from html.parser import HTMLParser as BaseHTMLParser + + class HTMLParser(BaseHTMLParser): + def __init__(self): + BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
Fix HTMLParser compatibility in Python 3
## Code Before: import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context ## Instruction: Fix HTMLParser compatibility in Python 3 ## Code After: import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
... if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) ...
cf8b49edfc38a98b4f6beba66bedcc13298eb114
yunity/utils/tests/mock.py
yunity/utils/tests/mock.py
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory from yunity.models import Category from yunity.utils.tests.fake import faker class Mock(DjangoModelFactory): class Meta: strategy = CREATE_STRATEGY model = None abstract = True class MockCategory(Mock): class Meta: model = "yunity.Category" strategy = CREATE_STRATEGY class MockUser(Mock): class Meta: model = "yunity.User" strategy = CREATE_STRATEGY is_active = True is_staff = False type = Category.objects.get(name='user.default') display_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = LazyAttribute(lambda _: faker.password()) locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)]) class MockChat(Mock): class Meta: model = "yunity.Chat" strategy = CREATE_STRATEGY administrated_by = SubFactory(MockUser) @post_generation def participants(self, create, extracted, **kwargs): if not create: return if extracted: for participant in extracted: self.participants.add(participant)
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory from yunity.models import Category from yunity.utils.tests.fake import faker class Mock(DjangoModelFactory): class Meta: strategy = CREATE_STRATEGY model = None abstract = True class MockCategory(Mock): class Meta: model = "yunity.Category" strategy = CREATE_STRATEGY class MockUser(Mock): class Meta: model = "yunity.User" strategy = CREATE_STRATEGY is_active = True is_staff = False type = Category.objects.get(name='user.default') display_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = LazyAttribute(lambda _: faker.password()) locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)]) class MockChat(Mock): class Meta: model = "yunity.Chat" strategy = CREATE_STRATEGY administrated_by = SubFactory(MockUser) @post_generation def participants(self, created, participants, **kwargs): if not created: return if participants: for participant in participants: self.participants.add(participant)
Rename some variables to try to explain magic
Rename some variables to try to explain magic
Python
agpl-3.0
yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/foodsaving-backend,yunity/yunity-core
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory from yunity.models import Category from yunity.utils.tests.fake import faker class Mock(DjangoModelFactory): class Meta: strategy = CREATE_STRATEGY model = None abstract = True class MockCategory(Mock): class Meta: model = "yunity.Category" strategy = CREATE_STRATEGY class MockUser(Mock): class Meta: model = "yunity.User" strategy = CREATE_STRATEGY is_active = True is_staff = False type = Category.objects.get(name='user.default') display_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = LazyAttribute(lambda _: faker.password()) locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)]) class MockChat(Mock): class Meta: model = "yunity.Chat" strategy = CREATE_STRATEGY administrated_by = SubFactory(MockUser) @post_generation - def participants(self, create, extracted, **kwargs): + def participants(self, created, participants, **kwargs): - if not create: + if not created: return - if extracted: + if participants: - for participant in extracted: + for participant in participants: self.participants.add(participant)
Rename some variables to try to explain magic
## Code Before: from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory from yunity.models import Category from yunity.utils.tests.fake import faker class Mock(DjangoModelFactory): class Meta: strategy = CREATE_STRATEGY model = None abstract = True class MockCategory(Mock): class Meta: model = "yunity.Category" strategy = CREATE_STRATEGY class MockUser(Mock): class Meta: model = "yunity.User" strategy = CREATE_STRATEGY is_active = True is_staff = False type = Category.objects.get(name='user.default') display_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = LazyAttribute(lambda _: faker.password()) locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)]) class MockChat(Mock): class Meta: model = "yunity.Chat" strategy = CREATE_STRATEGY administrated_by = SubFactory(MockUser) @post_generation def participants(self, create, extracted, **kwargs): if not create: return if extracted: for participant in extracted: self.participants.add(participant) ## Instruction: Rename some variables to try to explain magic ## Code After: from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory from yunity.models import Category from yunity.utils.tests.fake import faker class Mock(DjangoModelFactory): class Meta: strategy = CREATE_STRATEGY model = None abstract = True class MockCategory(Mock): class Meta: model = "yunity.Category" strategy = CREATE_STRATEGY class MockUser(Mock): class Meta: model = "yunity.User" strategy = CREATE_STRATEGY is_active = True is_staff = False type = Category.objects.get(name='user.default') display_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = LazyAttribute(lambda _: faker.password()) locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)]) class MockChat(Mock): class Meta: model = "yunity.Chat" strategy = CREATE_STRATEGY administrated_by = SubFactory(MockUser) @post_generation def participants(self, created, participants, **kwargs): if not created: return if participants: for participant in participants: self.participants.add(participant)
# ... existing code ... @post_generation def participants(self, created, participants, **kwargs): if not created: return if participants: for participant in participants: self.participants.add(participant) # ... rest of the code ...
f024e340a6a443bb765b67bbdb811fa44fd3d19b
tests/test_resources.py
tests/test_resources.py
from flask import json from helper import TestCase from models import db, Major class StudentsTestCase(TestCase): def setUp(self): super(StudentsTestCase, self).setUp() with self.appx.app_context(): db.session.add(Major(id=1, university_id=1, name='Major1')) db.session.add(Major(id=2, university_id=1, name='Major2')) db.session.commit() def test_students_patch(self): headers = { 'Authorization': 'Bearer ' + self.jwt, 'Content-Type': 'application/json' } data = { 'graduation_year': 2018, 'gender': 'm', 'majors': [1, 2] } rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data)) self.assertEqual(rv.status_code, 200)
from flask import json from helper import TestCase from models import db, Major, Student class StudentsTestCase(TestCase): def setUp(self): super(StudentsTestCase, self).setUp() with self.appx.app_context(): db.session.add(Major(id=1, university_id=1, name='Major1')) db.session.add(Major(id=2, university_id=1, name='Major2')) db.session.commit() def test_students_patch(self): headers = { 'Authorization': 'Bearer ' + self.jwt, 'Content-Type': 'application/json' } data = { 'graduation_year': 2018, 'gender': 'm', 'majors': [1, 2] } rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data)) self.assertEqual(rv.status_code, 200) with self.appx.app_context(): student = Student.query.get(0) self.assertEqual(student.graduation_year, data['graduation_year']) self.assertEqual(student.gender, data['gender']) self.assertEqual(student.majors_list, data['majors'])
Improve testing of student patching
Improve testing of student patching
Python
agpl-3.0
SCUEvals/scuevals-api,SCUEvals/scuevals-api
from flask import json from helper import TestCase - from models import db, Major + from models import db, Major, Student class StudentsTestCase(TestCase): def setUp(self): super(StudentsTestCase, self).setUp() with self.appx.app_context(): db.session.add(Major(id=1, university_id=1, name='Major1')) db.session.add(Major(id=2, university_id=1, name='Major2')) db.session.commit() def test_students_patch(self): headers = { 'Authorization': 'Bearer ' + self.jwt, 'Content-Type': 'application/json' } data = { 'graduation_year': 2018, 'gender': 'm', 'majors': [1, 2] } rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data)) self.assertEqual(rv.status_code, 200) + with self.appx.app_context(): + student = Student.query.get(0) + self.assertEqual(student.graduation_year, data['graduation_year']) + self.assertEqual(student.gender, data['gender']) + self.assertEqual(student.majors_list, data['majors']) +
Improve testing of student patching
## Code Before: from flask import json from helper import TestCase from models import db, Major class StudentsTestCase(TestCase): def setUp(self): super(StudentsTestCase, self).setUp() with self.appx.app_context(): db.session.add(Major(id=1, university_id=1, name='Major1')) db.session.add(Major(id=2, university_id=1, name='Major2')) db.session.commit() def test_students_patch(self): headers = { 'Authorization': 'Bearer ' + self.jwt, 'Content-Type': 'application/json' } data = { 'graduation_year': 2018, 'gender': 'm', 'majors': [1, 2] } rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data)) self.assertEqual(rv.status_code, 200) ## Instruction: Improve testing of student patching ## Code After: from flask import json from helper import TestCase from models import db, Major, Student class StudentsTestCase(TestCase): def setUp(self): super(StudentsTestCase, self).setUp() with self.appx.app_context(): db.session.add(Major(id=1, university_id=1, name='Major1')) db.session.add(Major(id=2, university_id=1, name='Major2')) db.session.commit() def test_students_patch(self): headers = { 'Authorization': 'Bearer ' + self.jwt, 'Content-Type': 'application/json' } data = { 'graduation_year': 2018, 'gender': 'm', 'majors': [1, 2] } rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data)) self.assertEqual(rv.status_code, 200) with self.appx.app_context(): student = Student.query.get(0) self.assertEqual(student.graduation_year, data['graduation_year']) self.assertEqual(student.gender, data['gender']) self.assertEqual(student.majors_list, data['majors'])
... from flask import json from helper import TestCase from models import db, Major, Student ... self.assertEqual(rv.status_code, 200) with self.appx.app_context(): student = Student.query.get(0) self.assertEqual(student.graduation_year, data['graduation_year']) self.assertEqual(student.gender, data['gender']) self.assertEqual(student.majors_list, data['majors']) ...
3d64eb4a7438b6b4f46f1fdf7f47d530cb11b09c
spacy/tests/regression/test_issue2396.py
spacy/tests/regression/test_issue2396.py
from __future__ import unicode_literals from ..util import get_doc import pytest import numpy @pytest.mark.parametrize('sentence,matrix', [ ( 'She created a test for spacy', numpy.array([ [0, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1], [1, 1, 2, 3, 3, 3], [1, 1, 3, 3, 3, 3], [1, 1, 3, 3, 4, 4], [1, 1, 3, 3, 4, 5]], dtype=numpy.int32) ) ]) def test_issue2396(EN, sentence, matrix): doc = EN(sentence) span = doc[:] assert (doc.get_lca_matrix() == matrix).all() assert (span.get_lca_matrix() == matrix).all()
from __future__ import unicode_literals from ..util import get_doc import pytest import numpy from numpy.testing import assert_array_equal @pytest.mark.parametrize('words,heads,matrix', [ ( 'She created a test for spacy'.split(), [1, 0, 1, -2, -1, -1], numpy.array([ [0, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1], [1, 1, 2, 3, 3, 3], [1, 1, 3, 3, 3, 3], [1, 1, 3, 3, 4, 4], [1, 1, 3, 3, 4, 5]], dtype=numpy.int32) ) ]) def test_issue2396(en_vocab, words, heads, matrix): doc = get_doc(en_vocab, words=words, heads=heads) span = doc[:] assert_array_equal(doc.get_lca_matrix(), matrix) assert_array_equal(span.get_lca_matrix(), matrix)
Update get_lca_matrix test for develop
Update get_lca_matrix test for develop
Python
mit
explosion/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,honnibal/spaCy,honnibal/spaCy,honnibal/spaCy,spacy-io/spaCy,honnibal/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy
from __future__ import unicode_literals from ..util import get_doc import pytest import numpy + from numpy.testing import assert_array_equal + - @pytest.mark.parametrize('sentence,matrix', [ + @pytest.mark.parametrize('words,heads,matrix', [ ( - 'She created a test for spacy', + 'She created a test for spacy'.split(), + [1, 0, 1, -2, -1, -1], numpy.array([ [0, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1], [1, 1, 2, 3, 3, 3], [1, 1, 3, 3, 3, 3], [1, 1, 3, 3, 4, 4], [1, 1, 3, 3, 4, 5]], dtype=numpy.int32) ) ]) - def test_issue2396(EN, sentence, matrix): - doc = EN(sentence) + def test_issue2396(en_vocab, words, heads, matrix): + doc = get_doc(en_vocab, words=words, heads=heads) + span = doc[:] - assert (doc.get_lca_matrix() == matrix).all() + assert_array_equal(doc.get_lca_matrix(), matrix) - assert (span.get_lca_matrix() == matrix).all() + assert_array_equal(span.get_lca_matrix(), matrix)
Update get_lca_matrix test for develop
## Code Before: from __future__ import unicode_literals from ..util import get_doc import pytest import numpy @pytest.mark.parametrize('sentence,matrix', [ ( 'She created a test for spacy', numpy.array([ [0, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1], [1, 1, 2, 3, 3, 3], [1, 1, 3, 3, 3, 3], [1, 1, 3, 3, 4, 4], [1, 1, 3, 3, 4, 5]], dtype=numpy.int32) ) ]) def test_issue2396(EN, sentence, matrix): doc = EN(sentence) span = doc[:] assert (doc.get_lca_matrix() == matrix).all() assert (span.get_lca_matrix() == matrix).all() ## Instruction: Update get_lca_matrix test for develop ## Code After: from __future__ import unicode_literals from ..util import get_doc import pytest import numpy from numpy.testing import assert_array_equal @pytest.mark.parametrize('words,heads,matrix', [ ( 'She created a test for spacy'.split(), [1, 0, 1, -2, -1, -1], numpy.array([ [0, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1], [1, 1, 2, 3, 3, 3], [1, 1, 3, 3, 3, 3], [1, 1, 3, 3, 4, 4], [1, 1, 3, 3, 4, 5]], dtype=numpy.int32) ) ]) def test_issue2396(en_vocab, words, heads, matrix): doc = get_doc(en_vocab, words=words, heads=heads) span = doc[:] assert_array_equal(doc.get_lca_matrix(), matrix) assert_array_equal(span.get_lca_matrix(), matrix)
# ... existing code ... import pytest import numpy from numpy.testing import assert_array_equal @pytest.mark.parametrize('words,heads,matrix', [ ( 'She created a test for spacy'.split(), [1, 0, 1, -2, -1, -1], numpy.array([ [0, 1, 1, 1, 1, 1], # ... modified code ... ) ]) def test_issue2396(en_vocab, words, heads, matrix): doc = get_doc(en_vocab, words=words, heads=heads) span = doc[:] assert_array_equal(doc.get_lca_matrix(), matrix) assert_array_equal(span.get_lca_matrix(), matrix) # ... rest of the code ...
210c7b7fb421a7c083b9d292370b15c0ece17fa7
source/bark/__init__.py
source/bark/__init__.py
from .handler.distribute import Distribute #: Top level handler responsible for relaying all logs to other handlers. handle = Distribute()
from .handler.distribute import Distribute #: Top level handler responsible for relaying all logs to other handlers. handler = Distribute() handlers = handler.handlers handle = handler.handle
Correct handler reference variable name and add convenient accessors.
Correct handler reference variable name and add convenient accessors.
Python
apache-2.0
4degrees/mill,4degrees/sawmill
from .handler.distribute import Distribute #: Top level handler responsible for relaying all logs to other handlers. - handle = Distribute() + handler = Distribute() + handlers = handler.handlers + handle = handler.handle +
Correct handler reference variable name and add convenient accessors.
## Code Before: from .handler.distribute import Distribute #: Top level handler responsible for relaying all logs to other handlers. handle = Distribute() ## Instruction: Correct handler reference variable name and add convenient accessors. ## Code After: from .handler.distribute import Distribute #: Top level handler responsible for relaying all logs to other handlers. handler = Distribute() handlers = handler.handlers handle = handler.handle
... #: Top level handler responsible for relaying all logs to other handlers. handler = Distribute() handlers = handler.handlers handle = handler.handle ...
e2722385831a0930765d2c4bb78a582d41f4b64b
src/sentry/replays.py
src/sentry/replays.py
from __future__ import absolute_import import socket from httplib import HTTPConnection, HTTPSConnection from urllib import urlencode from urlparse import urlparse class Replayer(object): def __init__(self, url, method, data=None, headers=None): self.url = url self.method = method self.data = data self.headers = headers def replay(self): urlparts = urlparse(self.url) if urlparts.scheme == 'http': conn_cls = HTTPConnection elif urlparts.scheme == 'https': conn_cls = HTTPSConnection else: raise ValueError(self.url) data = self.data if isinstance(data, dict): data = urlencode(data) if urlparts.query: full_url = urlparts.path + '?' + urlparts.query else: full_url = urlparts.path conn = conn_cls(urlparts.netloc) try: conn.request(self.method, full_url, data, self.headers or {}) response = conn.getresponse() except socket.error as e: return { 'status': 'error', 'reason': str(e), } return { 'status': response.status, 'reason': response.reason, 'headers': response.getheaders(), 'body': response.read(), }
from __future__ import absolute_import import requests class Replayer(object): def __init__(self, url, method, data=None, headers=None): self.url = url self.method = method self.data = data self.headers = headers def replay(self): try: response = requests.request( self.method, self.url, data=self.data, headers=self.headers or {} ) except requests.RequestException as e: return { 'status': 'error', 'reason': str(e), } return { 'status': response.status_code, 'reason': response.reason, 'headers': response.headers, 'body': response.content, }
Use requests instead of httplib to do replay
Use requests instead of httplib to do replay
Python
bsd-3-clause
beeftornado/sentry,nicholasserra/sentry,Kryz/sentry,JackDanger/sentry,imankulov/sentry,JamesMura/sentry,zenefits/sentry,kevinlondon/sentry,mvaled/sentry,JamesMura/sentry,ifduyue/sentry,looker/sentry,daevaorn/sentry,fotinakis/sentry,gencer/sentry,looker/sentry,JackDanger/sentry,mvaled/sentry,Natim/sentry,beeftornado/sentry,korealerts1/sentry,imankulov/sentry,zenefits/sentry,jean/sentry,alexm92/sentry,fotinakis/sentry,daevaorn/sentry,beeftornado/sentry,BuildingLink/sentry,ngonzalvez/sentry,BayanGroup/sentry,mvaled/sentry,mitsuhiko/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,korealerts1/sentry,mvaled/sentry,gencer/sentry,imankulov/sentry,Kryz/sentry,looker/sentry,felixbuenemann/sentry,jean/sentry,mitsuhiko/sentry,fotinakis/sentry,ifduyue/sentry,mvaled/sentry,korealerts1/sentry,kevinlondon/sentry,Natim/sentry,alexm92/sentry,zenefits/sentry,zenefits/sentry,BayanGroup/sentry,jean/sentry,mvaled/sentry,ifduyue/sentry,JamesMura/sentry,kevinlondon/sentry,BuildingLink/sentry,BayanGroup/sentry,JamesMura/sentry,nicholasserra/sentry,felixbuenemann/sentry,Kryz/sentry,BuildingLink/sentry,BuildingLink/sentry,Natim/sentry,BuildingLink/sentry,jean/sentry,JackDanger/sentry,felixbuenemann/sentry,fotinakis/sentry,nicholasserra/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,alexm92/sentry,gencer/sentry,gencer/sentry,gencer/sentry,zenefits/sentry,JamesMura/sentry,ifduyue/sentry,jean/sentry,ifduyue/sentry
from __future__ import absolute_import + import requests - - import socket - - from httplib import HTTPConnection, HTTPSConnection - from urllib import urlencode - from urlparse import urlparse class Replayer(object): def __init__(self, url, method, data=None, headers=None): self.url = url self.method = method self.data = data self.headers = headers def replay(self): - urlparts = urlparse(self.url) - if urlparts.scheme == 'http': - conn_cls = HTTPConnection - elif urlparts.scheme == 'https': - conn_cls = HTTPSConnection - else: - raise ValueError(self.url) - - data = self.data - if isinstance(data, dict): - data = urlencode(data) - - if urlparts.query: - full_url = urlparts.path + '?' + urlparts.query - else: - full_url = urlparts.path - - conn = conn_cls(urlparts.netloc) try: - conn.request(self.method, full_url, data, self.headers or {}) - - response = conn.getresponse() - except socket.error as e: + response = requests.request( + self.method, + self.url, + data=self.data, + headers=self.headers or {} + ) + except requests.RequestException as e: return { 'status': 'error', 'reason': str(e), } return { - 'status': response.status, + 'status': response.status_code, 'reason': response.reason, - 'headers': response.getheaders(), + 'headers': response.headers, - 'body': response.read(), + 'body': response.content, }
Use requests instead of httplib to do replay
## Code Before: from __future__ import absolute_import import socket from httplib import HTTPConnection, HTTPSConnection from urllib import urlencode from urlparse import urlparse class Replayer(object): def __init__(self, url, method, data=None, headers=None): self.url = url self.method = method self.data = data self.headers = headers def replay(self): urlparts = urlparse(self.url) if urlparts.scheme == 'http': conn_cls = HTTPConnection elif urlparts.scheme == 'https': conn_cls = HTTPSConnection else: raise ValueError(self.url) data = self.data if isinstance(data, dict): data = urlencode(data) if urlparts.query: full_url = urlparts.path + '?' + urlparts.query else: full_url = urlparts.path conn = conn_cls(urlparts.netloc) try: conn.request(self.method, full_url, data, self.headers or {}) response = conn.getresponse() except socket.error as e: return { 'status': 'error', 'reason': str(e), } return { 'status': response.status, 'reason': response.reason, 'headers': response.getheaders(), 'body': response.read(), } ## Instruction: Use requests instead of httplib to do replay ## Code After: from __future__ import absolute_import import requests class Replayer(object): def __init__(self, url, method, data=None, headers=None): self.url = url self.method = method self.data = data self.headers = headers def replay(self): try: response = requests.request( self.method, self.url, data=self.data, headers=self.headers or {} ) except requests.RequestException as e: return { 'status': 'error', 'reason': str(e), } return { 'status': response.status_code, 'reason': response.reason, 'headers': response.headers, 'body': response.content, }
// ... existing code ... from __future__ import absolute_import import requests // ... modified code ... def replay(self): try: response = requests.request( self.method, self.url, data=self.data, headers=self.headers or {} ) except requests.RequestException as e: return { 'status': 'error', ... return { 'status': response.status_code, 'reason': response.reason, 'headers': response.headers, 'body': response.content, } // ... rest of the code ...
3289027d2cc5b07a83dca422bfc14114854618f8
kazoo/__init__.py
kazoo/__init__.py
import os from kazoo.zkclient import ZooKeeperClient __all__ = ['ZooKeeperClient'] # ZK C client likes to spew log info to STDERR. disable that unless an # env is present. def disable_zookeeper_log(): import zookeeper zookeeper.set_log_stream(open('/dev/null')) if not "KAZOO_LOG_ENABLED" in os.environ: disable_zookeeper_log() def patch_extras(): # workaround for http://code.google.com/p/gevent/issues/detail?id=112 # gevent isn't patching threading._sleep which causes problems # for Condition objects from gevent import sleep import threading threading._sleep = sleep if "KAZOO_TEST_GEVENT_PATCH" in os.environ: from gevent import monkey; monkey.patch_all() patch_extras()
import os from kazoo.zkclient import ZooKeeperClient from kazoo.client import KazooClient __all__ = ['ZooKeeperClient', 'KazooClient'] # ZK C client likes to spew log info to STDERR. disable that unless an # env is present. def disable_zookeeper_log(): import zookeeper zookeeper.set_log_stream(open('/dev/null')) if not "KAZOO_LOG_ENABLED" in os.environ: disable_zookeeper_log() def patch_extras(): # workaround for http://code.google.com/p/gevent/issues/detail?id=112 # gevent isn't patching threading._sleep which causes problems # for Condition objects from gevent import sleep import threading threading._sleep = sleep if "KAZOO_TEST_GEVENT_PATCH" in os.environ: from gevent import monkey; monkey.patch_all() patch_extras()
Add KazooClient to top-level module
Add KazooClient to top-level module
Python
apache-2.0
nimbusproject/kazoo
import os from kazoo.zkclient import ZooKeeperClient + from kazoo.client import KazooClient - __all__ = ['ZooKeeperClient'] + __all__ = ['ZooKeeperClient', 'KazooClient'] # ZK C client likes to spew log info to STDERR. disable that unless an # env is present. def disable_zookeeper_log(): import zookeeper zookeeper.set_log_stream(open('/dev/null')) if not "KAZOO_LOG_ENABLED" in os.environ: disable_zookeeper_log() def patch_extras(): # workaround for http://code.google.com/p/gevent/issues/detail?id=112 # gevent isn't patching threading._sleep which causes problems # for Condition objects from gevent import sleep import threading threading._sleep = sleep if "KAZOO_TEST_GEVENT_PATCH" in os.environ: from gevent import monkey; monkey.patch_all() patch_extras()
Add KazooClient to top-level module
## Code Before: import os from kazoo.zkclient import ZooKeeperClient __all__ = ['ZooKeeperClient'] # ZK C client likes to spew log info to STDERR. disable that unless an # env is present. def disable_zookeeper_log(): import zookeeper zookeeper.set_log_stream(open('/dev/null')) if not "KAZOO_LOG_ENABLED" in os.environ: disable_zookeeper_log() def patch_extras(): # workaround for http://code.google.com/p/gevent/issues/detail?id=112 # gevent isn't patching threading._sleep which causes problems # for Condition objects from gevent import sleep import threading threading._sleep = sleep if "KAZOO_TEST_GEVENT_PATCH" in os.environ: from gevent import monkey; monkey.patch_all() patch_extras() ## Instruction: Add KazooClient to top-level module ## Code After: import os from kazoo.zkclient import ZooKeeperClient from kazoo.client import KazooClient __all__ = ['ZooKeeperClient', 'KazooClient'] # ZK C client likes to spew log info to STDERR. disable that unless an # env is present. def disable_zookeeper_log(): import zookeeper zookeeper.set_log_stream(open('/dev/null')) if not "KAZOO_LOG_ENABLED" in os.environ: disable_zookeeper_log() def patch_extras(): # workaround for http://code.google.com/p/gevent/issues/detail?id=112 # gevent isn't patching threading._sleep which causes problems # for Condition objects from gevent import sleep import threading threading._sleep = sleep if "KAZOO_TEST_GEVENT_PATCH" in os.environ: from gevent import monkey; monkey.patch_all() patch_extras()
# ... existing code ... from kazoo.zkclient import ZooKeeperClient from kazoo.client import KazooClient __all__ = ['ZooKeeperClient', 'KazooClient'] # ... rest of the code ...
1639200e5700b1170a9d2312a32c7991ed5198b4
tests/basics/boundmeth1.py
tests/basics/boundmeth1.py
print(type(repr([].append))) class A: def f(self): return 0 def g(self, a): return a def h(self, a, b, c, d, e, f): return a + b + c + d + e + f # bound method with no extra args m = A().f print(m()) # bound method with 1 extra arg m = A().g print(m(1)) # bound method with lots of extra args m = A().h print(m(1, 2, 3, 4, 5, 6))
print(type(repr([].append))) class A: def f(self): return 0 def g(self, a): return a def h(self, a, b, c, d, e, f): return a + b + c + d + e + f # bound method with no extra args m = A().f print(m()) # bound method with 1 extra arg m = A().g print(m(1)) # bound method with lots of extra args m = A().h print(m(1, 2, 3, 4, 5, 6)) # can't assign attributes to a bound method try: A().f.x = 1 except AttributeError: print('AttributeError')
Add test for assignment of attribute to bound method.
tests/basics: Add test for assignment of attribute to bound method.
Python
mit
ryannathans/micropython,bvernoux/micropython,HenrikSolver/micropython,dmazzella/micropython,lowRISC/micropython,toolmacher/micropython,ryannathans/micropython,cwyark/micropython,deshipu/micropython,mhoffma/micropython,HenrikSolver/micropython,Peetz0r/micropython-esp32,Timmenem/micropython,MrSurly/micropython,tralamazza/micropython,alex-robbins/micropython,chrisdearman/micropython,adafruit/circuitpython,trezor/micropython,deshipu/micropython,adafruit/circuitpython,tobbad/micropython,dmazzella/micropython,PappaPeppar/micropython,puuu/micropython,MrSurly/micropython-esp32,blazewicz/micropython,MrSurly/micropython,selste/micropython,swegener/micropython,tralamazza/micropython,mhoffma/micropython,AriZuu/micropython,PappaPeppar/micropython,lowRISC/micropython,henriknelson/micropython,torwag/micropython,puuu/micropython,toolmacher/micropython,toolmacher/micropython,kerneltask/micropython,mhoffma/micropython,deshipu/micropython,HenrikSolver/micropython,Peetz0r/micropython-esp32,mhoffma/micropython,tobbad/micropython,pozetroninc/micropython,toolmacher/micropython,AriZuu/micropython,ryannathans/micropython,hiway/micropython,pozetroninc/micropython,mhoffma/micropython,swegener/micropython,swegener/micropython,blazewicz/micropython,trezor/micropython,selste/micropython,HenrikSolver/micropython,adafruit/micropython,SHA2017-badge/micropython-esp32,henriknelson/micropython,alex-robbins/micropython,PappaPeppar/micropython,oopy/micropython,MrSurly/micropython-esp32,adafruit/micropython,TDAbboud/micropython,adafruit/circuitpython,tobbad/micropython,infinnovation/micropython,infinnovation/micropython,alex-robbins/micropython,henriknelson/micropython,pfalcon/micropython,ryannathans/micropython,adafruit/circuitpython,adafruit/circuitpython,ryannathans/micropython,MrSurly/micropython-esp32,infinnovation/micropython,TDAbboud/micropython,pozetroninc/micropython,lowRISC/micropython,adafruit/micropython,MrSurly/micropython,hiway/micropython,Peetz0r/micropython-esp32,tobbad/micropython,MrSurly/micropython,micropython/micropython-esp32,tralamazza/micropython,pramasoul/micropython,Timmenem/micropython,pfalcon/micropython,micropython/micropython-esp32,bvernoux/micropython,henriknelson/micropython,chrisdearman/micropython,adafruit/circuitpython,Timmenem/micropython,torwag/micropython,micropython/micropython-esp32,AriZuu/micropython,dmazzella/micropython,pramasoul/micropython,selste/micropython,tralamazza/micropython,lowRISC/micropython,puuu/micropython,AriZuu/micropython,SHA2017-badge/micropython-esp32,alex-robbins/micropython,adafruit/micropython,HenrikSolver/micropython,pramasoul/micropython,kerneltask/micropython,cwyark/micropython,trezor/micropython,henriknelson/micropython,tobbad/micropython,hiway/micropython,Peetz0r/micropython-esp32,PappaPeppar/micropython,AriZuu/micropython,bvernoux/micropython,TDAbboud/micropython,swegener/micropython,adafruit/micropython,selste/micropython,SHA2017-badge/micropython-esp32,Timmenem/micropython,hiway/micropython,MrSurly/micropython,infinnovation/micropython,deshipu/micropython,pozetroninc/micropython,deshipu/micropython,MrSurly/micropython-esp32,micropython/micropython-esp32,MrSurly/micropython-esp32,trezor/micropython,trezor/micropython,SHA2017-badge/micropython-esp32,torwag/micropython,pramasoul/micropython,chrisdearman/micropython,infinnovation/micropython,blazewicz/micropython,TDAbboud/micropython,pramasoul/micropython,oopy/micropython,Peetz0r/micropython-esp32,micropython/micropython-esp32,pfalcon/micropython,kerneltask/micropython,TDAbboud/micropython,chrisdearman/micropython,torwag/micropython,blazewicz/micropython,pozetroninc/micropython,cwyark/micropython,pfalcon/micropython,kerneltask/micropython,cwyark/micropython,oopy/micropython,Timmenem/micropython,toolmacher/micropython,pfalcon/micropython,puuu/micropython,SHA2017-badge/micropython-esp32,puuu/micropython,bvernoux/micropython,blazewicz/micropython,oopy/micropython,chrisdearman/micropython,kerneltask/micropython,swegener/micropython,oopy/micropython,torwag/micropython,PappaPeppar/micropython,bvernoux/micropython,alex-robbins/micropython,hiway/micropython,cwyark/micropython,selste/micropython,lowRISC/micropython,dmazzella/micropython
print(type(repr([].append))) class A: def f(self): return 0 def g(self, a): return a def h(self, a, b, c, d, e, f): return a + b + c + d + e + f # bound method with no extra args m = A().f print(m()) # bound method with 1 extra arg m = A().g print(m(1)) # bound method with lots of extra args m = A().h print(m(1, 2, 3, 4, 5, 6)) + # can't assign attributes to a bound method + try: + A().f.x = 1 + except AttributeError: + print('AttributeError') +
Add test for assignment of attribute to bound method.
## Code Before: print(type(repr([].append))) class A: def f(self): return 0 def g(self, a): return a def h(self, a, b, c, d, e, f): return a + b + c + d + e + f # bound method with no extra args m = A().f print(m()) # bound method with 1 extra arg m = A().g print(m(1)) # bound method with lots of extra args m = A().h print(m(1, 2, 3, 4, 5, 6)) ## Instruction: Add test for assignment of attribute to bound method. ## Code After: print(type(repr([].append))) class A: def f(self): return 0 def g(self, a): return a def h(self, a, b, c, d, e, f): return a + b + c + d + e + f # bound method with no extra args m = A().f print(m()) # bound method with 1 extra arg m = A().g print(m(1)) # bound method with lots of extra args m = A().h print(m(1, 2, 3, 4, 5, 6)) # can't assign attributes to a bound method try: A().f.x = 1 except AttributeError: print('AttributeError')
// ... existing code ... m = A().h print(m(1, 2, 3, 4, 5, 6)) # can't assign attributes to a bound method try: A().f.x = 1 except AttributeError: print('AttributeError') // ... rest of the code ...
1e182ec0fd7cf550c809f2e6792629caeb8d5553
sauce/lib/helpers.py
sauce/lib/helpers.py
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) def striphtml(text): return re.sub('<[^<]+?>', ' ', text).strip()
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words from genshi.core import striptags import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs)
Use striptags from genshi for striphtml, since we have to have genshi anyway
Use striptags from genshi for striphtml, since we have to have genshi anyway
Python
agpl-3.0
moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words + from genshi.core import striptags + import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) + striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) - def striphtml(text): - return re.sub('<[^<]+?>', ' ', text).strip()
Use striptags from genshi for striphtml, since we have to have genshi anyway
## Code Before: from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) def striphtml(text): return re.sub('<[^<]+?>', ' ', text).strip() ## Instruction: Use striptags from genshi for striphtml, since we have to have genshi anyway ## Code After: from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words from genshi.core import striptags import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs)
// ... existing code ... from webhelpers.date import distance_of_time_in_words from genshi.core import striptags import re // ... modified code ... cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags def link(label, url='', **attrs): ... return link_to(label, tgurl(url), **attrs) // ... rest of the code ...
2ab601492a76be5d32a2e1d5009c150269e5fb03
src/interviews/managers.py
src/interviews/managers.py
import logging from datetime import timedelta from django.db import models from django.utils import timezone from .google_analytics import get_most_read_pages logger = logging.getLogger(__name__) class InterviewManager(models.Manager): def active(self, *args, **kwargs): return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now()) def newest(self, *args, **kwargs): return self.active().first() def last_week(self, *args, **kwargs): some_day_last_week = timezone.now().date() - timedelta(days=7) monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1)) monday_of_this_week = monday_of_last_week + timedelta(days=7) return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week, publish__lt=monday_of_this_week)[:1] def most_read(self, *args, **kwargs): slugs = get_most_read_pages() if slugs: return self.active().filter(slug__in=slugs) return super(InterviewManager, self).none()
import logging from datetime import timedelta from django.db import models from django.utils import timezone from .google_analytics import get_most_read_pages logger = logging.getLogger(__name__) class InterviewManager(models.Manager): def active(self, *args, **kwargs): return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now()) def newest(self, *args, **kwargs): return self.active().first() def last_week(self, *args, **kwargs): some_day_last_week = timezone.now().date() - timedelta(days=7) monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1)) monday_of_this_week = monday_of_last_week + timedelta(days=7) return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week, publish__lt=monday_of_this_week)[:1] def most_read(self, *args, **kwargs): slugs = get_most_read_pages() if slugs: return self.active().filter(slug__in=slugs).order_by('slug') return super(InterviewManager, self).none()
Order `most_read` queryset by slug.
Order `most_read` queryset by slug.
Python
mit
vermpy/thespotlight,vermpy/thespotlight,vermpy/thespotlight
import logging from datetime import timedelta from django.db import models from django.utils import timezone from .google_analytics import get_most_read_pages logger = logging.getLogger(__name__) class InterviewManager(models.Manager): def active(self, *args, **kwargs): return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now()) def newest(self, *args, **kwargs): return self.active().first() def last_week(self, *args, **kwargs): some_day_last_week = timezone.now().date() - timedelta(days=7) monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1)) monday_of_this_week = monday_of_last_week + timedelta(days=7) return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week, publish__lt=monday_of_this_week)[:1] def most_read(self, *args, **kwargs): slugs = get_most_read_pages() if slugs: - return self.active().filter(slug__in=slugs) + return self.active().filter(slug__in=slugs).order_by('slug') return super(InterviewManager, self).none()
Order `most_read` queryset by slug.
## Code Before: import logging from datetime import timedelta from django.db import models from django.utils import timezone from .google_analytics import get_most_read_pages logger = logging.getLogger(__name__) class InterviewManager(models.Manager): def active(self, *args, **kwargs): return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now()) def newest(self, *args, **kwargs): return self.active().first() def last_week(self, *args, **kwargs): some_day_last_week = timezone.now().date() - timedelta(days=7) monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1)) monday_of_this_week = monday_of_last_week + timedelta(days=7) return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week, publish__lt=monday_of_this_week)[:1] def most_read(self, *args, **kwargs): slugs = get_most_read_pages() if slugs: return self.active().filter(slug__in=slugs) return super(InterviewManager, self).none() ## Instruction: Order `most_read` queryset by slug. ## Code After: import logging from datetime import timedelta from django.db import models from django.utils import timezone from .google_analytics import get_most_read_pages logger = logging.getLogger(__name__) class InterviewManager(models.Manager): def active(self, *args, **kwargs): return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now()) def newest(self, *args, **kwargs): return self.active().first() def last_week(self, *args, **kwargs): some_day_last_week = timezone.now().date() - timedelta(days=7) monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1)) monday_of_this_week = monday_of_last_week + timedelta(days=7) return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week, publish__lt=monday_of_this_week)[:1] def most_read(self, *args, **kwargs): slugs = get_most_read_pages() if slugs: return self.active().filter(slug__in=slugs).order_by('slug') return super(InterviewManager, self).none()
// ... existing code ... slugs = get_most_read_pages() if slugs: return self.active().filter(slug__in=slugs).order_by('slug') return super(InterviewManager, self).none() // ... rest of the code ...
c37e0b66b6f0cc57d7df94f62dd47e00dc91c544
django_archive/archivers/__init__.py
django_archive/archivers/__init__.py
from .tarball import TarballArchiver from .zipfile import ZipArchiver TARBALL = TarballArchiver.UNCOMPRESSED TARBALL_GZ = TarballArchiver.GZ TARBALL_BZ2 = TarballArchiver.BZ2 TARBALL_XZ = TarballArchiver.XZ ZIP = 'zip' FORMATS = ( (TARBALL, "Tarball (.tar)"), (TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"), (TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"), (TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"), (ZIP, "ZIP archive (.zip)"), ) def get_archiver(fmt): """ Return the class corresponding with the provided archival format """ if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ): return TarballArchiver if fmt == ZIP: return ZipArchiver raise KeyError("Invalid format '{}' specified".format(fmt))
from .tarball import TarballArchiver from .zipfile import ZipArchiver TARBALL = TarballArchiver.UNCOMPRESSED TARBALL_GZ = TarballArchiver.GZ TARBALL_BZ2 = TarballArchiver.BZ2 TARBALL_XZ = TarballArchiver.XZ ZIP = 'zip' FORMATS = ( TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ, ZIP, ) FORMATS_DESC = { TARBALL: "Tarball (.tar)", TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)", TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)", TARBALL_XZ: "xz-compressed Tarball (.tar.xz)", ZIP: "ZIP archive (.zip)", } def get_archiver(fmt): """ Return the class corresponding with the provided archival format """ if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ): return TarballArchiver if fmt == ZIP: return ZipArchiver raise KeyError("Invalid format '{}' specified".format(fmt))
Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions.
Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions.
Python
mit
nathan-osman/django-archive,nathan-osman/django-archive
from .tarball import TarballArchiver from .zipfile import ZipArchiver TARBALL = TarballArchiver.UNCOMPRESSED TARBALL_GZ = TarballArchiver.GZ TARBALL_BZ2 = TarballArchiver.BZ2 TARBALL_XZ = TarballArchiver.XZ ZIP = 'zip' FORMATS = ( - (TARBALL, "Tarball (.tar)"), - (TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"), - (TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"), - (TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"), - (ZIP, "ZIP archive (.zip)"), + TARBALL, + TARBALL_GZ, + TARBALL_BZ2, + TARBALL_XZ, + ZIP, ) + + FORMATS_DESC = { + TARBALL: "Tarball (.tar)", + TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)", + TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)", + TARBALL_XZ: "xz-compressed Tarball (.tar.xz)", + ZIP: "ZIP archive (.zip)", + } def get_archiver(fmt): """ Return the class corresponding with the provided archival format """ if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ): return TarballArchiver if fmt == ZIP: return ZipArchiver raise KeyError("Invalid format '{}' specified".format(fmt))
Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions.
## Code Before: from .tarball import TarballArchiver from .zipfile import ZipArchiver TARBALL = TarballArchiver.UNCOMPRESSED TARBALL_GZ = TarballArchiver.GZ TARBALL_BZ2 = TarballArchiver.BZ2 TARBALL_XZ = TarballArchiver.XZ ZIP = 'zip' FORMATS = ( (TARBALL, "Tarball (.tar)"), (TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"), (TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"), (TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"), (ZIP, "ZIP archive (.zip)"), ) def get_archiver(fmt): """ Return the class corresponding with the provided archival format """ if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ): return TarballArchiver if fmt == ZIP: return ZipArchiver raise KeyError("Invalid format '{}' specified".format(fmt)) ## Instruction: Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions. ## Code After: from .tarball import TarballArchiver from .zipfile import ZipArchiver TARBALL = TarballArchiver.UNCOMPRESSED TARBALL_GZ = TarballArchiver.GZ TARBALL_BZ2 = TarballArchiver.BZ2 TARBALL_XZ = TarballArchiver.XZ ZIP = 'zip' FORMATS = ( TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ, ZIP, ) FORMATS_DESC = { TARBALL: "Tarball (.tar)", TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)", TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)", TARBALL_XZ: "xz-compressed Tarball (.tar.xz)", ZIP: "ZIP archive (.zip)", } def get_archiver(fmt): """ Return the class corresponding with the provided archival format """ if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ): return TarballArchiver if fmt == ZIP: return ZipArchiver raise KeyError("Invalid format '{}' specified".format(fmt))
// ... existing code ... FORMATS = ( TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ, ZIP, ) FORMATS_DESC = { TARBALL: "Tarball (.tar)", TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)", TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)", TARBALL_XZ: "xz-compressed Tarball (.tar.xz)", ZIP: "ZIP archive (.zip)", } // ... rest of the code ...
4078743923befac99672b67ea53fd1fe11af2e8c
tests/test_mjviewer.py
tests/test_mjviewer.py
import unittest from mujoco_py import mjviewer, mjcore class MjLibTest(unittest.TestCase): xml_path = 'tests/models/cartpole.xml' def setUp(self): self.width = 100 self.height = 100 self.viewer = mjviewer.MjViewer(visible=False, init_width=self.width, init_height=self.height) def tearDown(self): self.viewer.finish() self.viewer = None def test_start(self): self.viewer.start() self.assertTrue(self.viewer.running) def test_render(self): self.viewer.start() model = mjcore.MjModel(self.xml_path) self.viewer.set_model(model) (data, width, height) = self.viewer.get_image() # check image size is consistent # note that width and height may not equal self.width and self.height # e.g. on a computer with retina screen, # the width and height are scaled self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black self.assertTrue(any(map(ord, data)))
import unittest from mujoco_py import mjviewer, mjcore class MjLibTest(unittest.TestCase): xml_path = 'tests/models/cartpole.xml' def setUp(self): self.width = 100 self.height = 100 self.viewer = mjviewer.MjViewer(visible=False, init_width=self.width, init_height=self.height) def tearDown(self): self.viewer.finish() self.viewer = None def test_start(self): self.viewer.start() self.assertTrue(self.viewer.running) def test_render(self): self.viewer.start() model = mjcore.MjModel(self.xml_path) self.viewer.set_model(model) (data, width, height) = self.viewer.get_image() # check image size is consistent # note that width and height may not equal self.width and self.height # e.g. on a computer with retina screen, # the width and height are scaled self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black self.assertTrue(any(map(lambda x: x > 0, data)))
Stop using ord with ints
Stop using ord with ints
Python
mit
pulkitag/mujoco140-py,pulkitag/mujoco140-py,pulkitag/mujoco140-py
import unittest from mujoco_py import mjviewer, mjcore class MjLibTest(unittest.TestCase): xml_path = 'tests/models/cartpole.xml' def setUp(self): self.width = 100 self.height = 100 self.viewer = mjviewer.MjViewer(visible=False, init_width=self.width, init_height=self.height) def tearDown(self): self.viewer.finish() self.viewer = None def test_start(self): self.viewer.start() self.assertTrue(self.viewer.running) def test_render(self): self.viewer.start() model = mjcore.MjModel(self.xml_path) self.viewer.set_model(model) (data, width, height) = self.viewer.get_image() # check image size is consistent # note that width and height may not equal self.width and self.height # e.g. on a computer with retina screen, # the width and height are scaled self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black - self.assertTrue(any(map(ord, data))) + self.assertTrue(any(map(lambda x: x > 0, data)))
Stop using ord with ints
## Code Before: import unittest from mujoco_py import mjviewer, mjcore class MjLibTest(unittest.TestCase): xml_path = 'tests/models/cartpole.xml' def setUp(self): self.width = 100 self.height = 100 self.viewer = mjviewer.MjViewer(visible=False, init_width=self.width, init_height=self.height) def tearDown(self): self.viewer.finish() self.viewer = None def test_start(self): self.viewer.start() self.assertTrue(self.viewer.running) def test_render(self): self.viewer.start() model = mjcore.MjModel(self.xml_path) self.viewer.set_model(model) (data, width, height) = self.viewer.get_image() # check image size is consistent # note that width and height may not equal self.width and self.height # e.g. on a computer with retina screen, # the width and height are scaled self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black self.assertTrue(any(map(ord, data))) ## Instruction: Stop using ord with ints ## Code After: import unittest from mujoco_py import mjviewer, mjcore class MjLibTest(unittest.TestCase): xml_path = 'tests/models/cartpole.xml' def setUp(self): self.width = 100 self.height = 100 self.viewer = mjviewer.MjViewer(visible=False, init_width=self.width, init_height=self.height) def tearDown(self): self.viewer.finish() self.viewer = None def test_start(self): self.viewer.start() self.assertTrue(self.viewer.running) def test_render(self): self.viewer.start() model = mjcore.MjModel(self.xml_path) self.viewer.set_model(model) (data, width, height) = self.viewer.get_image() # check image size is consistent # note that width and height may not equal self.width and self.height # e.g. on a computer with retina screen, # the width and height are scaled self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black self.assertTrue(any(map(lambda x: x > 0, data)))
... self.assertEqual(len(data), 3 * width * height) # make sure the image is not pitch black self.assertTrue(any(map(lambda x: x > 0, data))) ...
6ae83f01eacceb140435e72a216fa88bd97f2b0c
pyswarms/utils/console_utils.py
pyswarms/utils/console_utils.py
""" console_utils.py: various tools for printing into console """ def cli_print(message, verbosity, threshold): """Helper function to print console output Parameters ---------- message : str the message to be printed into the console verbosity : int verbosity setting of the user threshold : int threshold for printing """ if verbosity >= threshold: print(message) else: pass def end_report(cost, pos, verbosity): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. Parameters ---------- cost : float final cost from the optimization procedure. pos : numpy.ndarray or list best position found verbosity : int verbosity setting of the user. """ # Cuts the length of the best position if it's too long if len(list(pos)) > 3: out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos)) else: out = list(pos) template = ('================================\n' 'Optimization finished!\n' 'Final cost: {:06.4f}\n' 'Best value: {}\n').format(cost, out) if verbosity >= 1: print(template)
""" console_utils.py: various tools for printing into console """ # Import from __future__ from __future__ import with_statement from __future__ import absolute_import from __future__ import print_function # Import modules import logging def cli_print(message, verbosity, threshold, logger): """Helper function to print console output Parameters ---------- message : str the message to be printed into the console verbosity : int verbosity setting of the user threshold : int threshold for printing logger : logging.getLogger logger instance """ if verbosity >= threshold: logger.info(message) else: pass def end_report(cost, pos, verbosity, logger): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. Parameters ---------- cost : float final cost from the optimization procedure. pos : numpy.ndarray or list best position found verbosity : int verbosity setting of the user. logger : logging.getLogger logger instance """ # Cuts the length of the best position if it's too long if len(list(pos)) > 3: out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos)) else: out = list(pos) template = ('================================\n' 'Optimization finished!\n' 'Final cost: {:06.4f}\n' 'Best value: {}\n').format(cost, out) if verbosity >= 1: logger.info(template)
Add support for logging module
Add support for logging module This package now prints using the logging module. It can still print onto the console, but an additional tag like INFO, DEBUG, etc. are now being used. Author: ljvmiranda921
Python
mit
ljvmiranda921/pyswarms,ljvmiranda921/pyswarms
""" console_utils.py: various tools for printing into console """ + # Import from __future__ + from __future__ import with_statement + from __future__ import absolute_import + from __future__ import print_function + + # Import modules + import logging + - def cli_print(message, verbosity, threshold): + def cli_print(message, verbosity, threshold, logger): """Helper function to print console output Parameters ---------- message : str the message to be printed into the console verbosity : int verbosity setting of the user threshold : int threshold for printing + logger : logging.getLogger + logger instance """ if verbosity >= threshold: - print(message) + logger.info(message) else: pass - def end_report(cost, pos, verbosity): + def end_report(cost, pos, verbosity, logger): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. Parameters ---------- cost : float final cost from the optimization procedure. pos : numpy.ndarray or list best position found verbosity : int verbosity setting of the user. + logger : logging.getLogger + logger instance """ # Cuts the length of the best position if it's too long if len(list(pos)) > 3: out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos)) else: out = list(pos) template = ('================================\n' 'Optimization finished!\n' 'Final cost: {:06.4f}\n' 'Best value: {}\n').format(cost, out) if verbosity >= 1: - print(template) + logger.info(template)
Add support for logging module
## Code Before: """ console_utils.py: various tools for printing into console """ def cli_print(message, verbosity, threshold): """Helper function to print console output Parameters ---------- message : str the message to be printed into the console verbosity : int verbosity setting of the user threshold : int threshold for printing """ if verbosity >= threshold: print(message) else: pass def end_report(cost, pos, verbosity): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. Parameters ---------- cost : float final cost from the optimization procedure. pos : numpy.ndarray or list best position found verbosity : int verbosity setting of the user. """ # Cuts the length of the best position if it's too long if len(list(pos)) > 3: out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos)) else: out = list(pos) template = ('================================\n' 'Optimization finished!\n' 'Final cost: {:06.4f}\n' 'Best value: {}\n').format(cost, out) if verbosity >= 1: print(template) ## Instruction: Add support for logging module ## Code After: """ console_utils.py: various tools for printing into console """ # Import from __future__ from __future__ import with_statement from __future__ import absolute_import from __future__ import print_function # Import modules import logging def cli_print(message, verbosity, threshold, logger): """Helper function to print console output Parameters ---------- message : str the message to be printed into the console verbosity : int verbosity setting of the user threshold : int threshold for printing logger : logging.getLogger logger instance """ if verbosity >= threshold: logger.info(message) else: pass def end_report(cost, pos, verbosity, logger): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. Parameters ---------- cost : float final cost from the optimization procedure. pos : numpy.ndarray or list best position found verbosity : int verbosity setting of the user. logger : logging.getLogger logger instance """ # Cuts the length of the best position if it's too long if len(list(pos)) > 3: out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos)) else: out = list(pos) template = ('================================\n' 'Optimization finished!\n' 'Final cost: {:06.4f}\n' 'Best value: {}\n').format(cost, out) if verbosity >= 1: logger.info(template)
# ... existing code ... """ console_utils.py: various tools for printing into console """ # Import from __future__ from __future__ import with_statement from __future__ import absolute_import from __future__ import print_function # Import modules import logging def cli_print(message, verbosity, threshold, logger): """Helper function to print console output # ... modified code ... threshold : int threshold for printing logger : logging.getLogger logger instance """ if verbosity >= threshold: logger.info(message) else: pass def end_report(cost, pos, verbosity, logger): """Helper function to print a simple report at the end of the run. This always has a threshold of 1. ... verbosity : int verbosity setting of the user. logger : logging.getLogger logger instance """ ... 'Best value: {}\n').format(cost, out) if verbosity >= 1: logger.info(template) # ... rest of the code ...
7ed12facca2f94eb8bba721e9b11882ea24726fe
crmapp/subscribers/views.py
crmapp/subscribers/views.py
from django.shortcuts import render from django.contrib.auth.models import User from django.http import HttpResponseRedirect from .forms import SubscriberForm def subscriber_new(request, template='subscribers/subscriber_new.html'): if request.method == 'POST': form = SubscriberForm(request.POST) if form.is_valid(): # Unpack form values username = form.cleaned_data['username'] password = form.cleaned_data['password1'] email = form.cleaned_data['email'] # Create the User record user = User(username=username, email=email) user.set_password(password) user.save() # Create Subscriber Record # Process payment (via Stripe) # Auto login the user return HttpResponseRedirect('/success/') else: form = SubscriberForm() return render(request, template, {'form':form})
from django.shortcuts import render from django.contrib.auth.models import User from django.http import HttpResponseRedirect from .forms import SubscriberForm from .models import Subscriber def subscriber_new(request, template='subscribers/subscriber_new.html'): if request.method == 'POST': form = SubscriberForm(request.POST) if form.is_valid(): # Unpack form values username = form.cleaned_data['username'] password = form.cleaned_data['password1'] email = form.cleaned_data['email'] first_name = form.cleaned_data['first_name'] last_name = form.cleaned_data['last_name'] # Create the User record user = User(username=username, email=email, first_name=first_name, last_name=last_name) user.set_password(password) user.save() # Create Subscriber Record address_one = form.cleaned_data['address_one'] address_two = form.cleaned_data['address_two'] city = form.cleaned_data['city'] state = form.cleaned_data['state'] sub = Subscriber(address_one=address_one, address_two=address_two, city=city, state=state, user_rec=user) sub.save() # Process payment (via Stripe) # Auto login the user return HttpResponseRedirect('/success/') else: form = SubscriberForm() return render(request, template, {'form':form})
Create the Subscriber Form - Part II > Update the View
Create the Subscriber Form - Part II > Update the View
Python
mit
deenaariff/Django,tabdon/crmeasyapp,tabdon/crmeasyapp
from django.shortcuts import render from django.contrib.auth.models import User from django.http import HttpResponseRedirect from .forms import SubscriberForm + from .models import Subscriber def subscriber_new(request, template='subscribers/subscriber_new.html'): if request.method == 'POST': form = SubscriberForm(request.POST) if form.is_valid(): # Unpack form values username = form.cleaned_data['username'] password = form.cleaned_data['password1'] email = form.cleaned_data['email'] + first_name = form.cleaned_data['first_name'] + last_name = form.cleaned_data['last_name'] # Create the User record - user = User(username=username, email=email) + user = User(username=username, email=email, + first_name=first_name, last_name=last_name) user.set_password(password) user.save() # Create Subscriber Record + address_one = form.cleaned_data['address_one'] + address_two = form.cleaned_data['address_two'] + city = form.cleaned_data['city'] + state = form.cleaned_data['state'] + sub = Subscriber(address_one=address_one, address_two=address_two, + city=city, state=state, user_rec=user) + sub.save() # Process payment (via Stripe) # Auto login the user return HttpResponseRedirect('/success/') else: form = SubscriberForm() return render(request, template, {'form':form})
Create the Subscriber Form - Part II > Update the View
## Code Before: from django.shortcuts import render from django.contrib.auth.models import User from django.http import HttpResponseRedirect from .forms import SubscriberForm def subscriber_new(request, template='subscribers/subscriber_new.html'): if request.method == 'POST': form = SubscriberForm(request.POST) if form.is_valid(): # Unpack form values username = form.cleaned_data['username'] password = form.cleaned_data['password1'] email = form.cleaned_data['email'] # Create the User record user = User(username=username, email=email) user.set_password(password) user.save() # Create Subscriber Record # Process payment (via Stripe) # Auto login the user return HttpResponseRedirect('/success/') else: form = SubscriberForm() return render(request, template, {'form':form}) ## Instruction: Create the Subscriber Form - Part II > Update the View ## Code After: from django.shortcuts import render from django.contrib.auth.models import User from django.http import HttpResponseRedirect from .forms import SubscriberForm from .models import Subscriber def subscriber_new(request, template='subscribers/subscriber_new.html'): if request.method == 'POST': form = SubscriberForm(request.POST) if form.is_valid(): # Unpack form values username = form.cleaned_data['username'] password = form.cleaned_data['password1'] email = form.cleaned_data['email'] first_name = form.cleaned_data['first_name'] last_name = form.cleaned_data['last_name'] # Create the User record user = User(username=username, email=email, first_name=first_name, last_name=last_name) user.set_password(password) user.save() # Create Subscriber Record address_one = form.cleaned_data['address_one'] address_two = form.cleaned_data['address_two'] city = form.cleaned_data['city'] state = form.cleaned_data['state'] sub = Subscriber(address_one=address_one, address_two=address_two, city=city, state=state, user_rec=user) sub.save() # Process payment (via Stripe) # Auto login the user return HttpResponseRedirect('/success/') else: form = SubscriberForm() return render(request, template, {'form':form})
// ... existing code ... from .forms import SubscriberForm from .models import Subscriber def subscriber_new(request, template='subscribers/subscriber_new.html'): // ... modified code ... password = form.cleaned_data['password1'] email = form.cleaned_data['email'] first_name = form.cleaned_data['first_name'] last_name = form.cleaned_data['last_name'] # Create the User record user = User(username=username, email=email, first_name=first_name, last_name=last_name) user.set_password(password) user.save() # Create Subscriber Record address_one = form.cleaned_data['address_one'] address_two = form.cleaned_data['address_two'] city = form.cleaned_data['city'] state = form.cleaned_data['state'] sub = Subscriber(address_one=address_one, address_two=address_two, city=city, state=state, user_rec=user) sub.save() # Process payment (via Stripe) # Auto login the user // ... rest of the code ...
e49e7484987e3b508802adbd9e05b2b156eb6bdd
manage.py
manage.py
import os import coverage from flask_script import Manager, Shell from flask_migrate import Migrate, MigrateCommand from config import basedir from app import create_app, db from app.models import User, Dictionary app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") migrate = Migrate(app, db) manager = Manager(app) def make_shell_context(): return dict(app=app, db=db, User=User, Dictionary=Dictionary) manager.add_command("shell", Shell(make_context=make_shell_context)) manager.add_command("db", MigrateCommand) cov = coverage.coverage(branch=True, include="app/*") @manager.command def test(coverage=False): """ Run the unit tests. """ if coverage: cov.start() import unittest tests = unittest.TestLoader().discover("tests") unittest.TextTestRunner(verbosity=2).run(tests) if coverage: cov.stop() cov.save() print("Coverage Summary:") cov.report() cov_dir = os.path.join(basedir, "tmp/coverage") cov.html_report(directory=cov_dir) print("HTML version: %s/index.html" % cov_dir) cov.erase() if __name__ == "__main__": manager.run()
import os import coverage from flask_script import Manager, Shell from flask_migrate import Migrate, MigrateCommand from config import basedir from app import create_app, db from app.models import User, Dictionary, Word app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") migrate = Migrate(app, db) manager = Manager(app) def make_shell_context(): return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word) manager.add_command("shell", Shell(make_context=make_shell_context)) manager.add_command("db", MigrateCommand) cov = coverage.coverage(branch=True, include="app/*") @manager.command def test(coverage=False): """ Run the unit tests. """ if coverage: cov.start() import unittest tests = unittest.TestLoader().discover("tests") unittest.TextTestRunner(verbosity=2).run(tests) if coverage: cov.stop() cov.save() print("Coverage Summary:") cov.report() cov_dir = os.path.join(basedir, "tmp/coverage") cov.html_report(directory=cov_dir) print("HTML version: %s/index.html" % cov_dir) cov.erase() if __name__ == "__main__": manager.run()
Add Word model to shell context
Add Word model to shell context
Python
mit
Encrylize/MyDictionary,Encrylize/MyDictionary,Encrylize/MyDictionary
import os import coverage from flask_script import Manager, Shell from flask_migrate import Migrate, MigrateCommand from config import basedir from app import create_app, db - from app.models import User, Dictionary + from app.models import User, Dictionary, Word app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") migrate = Migrate(app, db) manager = Manager(app) def make_shell_context(): - return dict(app=app, db=db, User=User, Dictionary=Dictionary) + return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word) manager.add_command("shell", Shell(make_context=make_shell_context)) manager.add_command("db", MigrateCommand) cov = coverage.coverage(branch=True, include="app/*") @manager.command def test(coverage=False): """ Run the unit tests. """ if coverage: cov.start() import unittest tests = unittest.TestLoader().discover("tests") unittest.TextTestRunner(verbosity=2).run(tests) if coverage: cov.stop() cov.save() print("Coverage Summary:") cov.report() cov_dir = os.path.join(basedir, "tmp/coverage") cov.html_report(directory=cov_dir) print("HTML version: %s/index.html" % cov_dir) cov.erase() if __name__ == "__main__": manager.run()
Add Word model to shell context
## Code Before: import os import coverage from flask_script import Manager, Shell from flask_migrate import Migrate, MigrateCommand from config import basedir from app import create_app, db from app.models import User, Dictionary app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") migrate = Migrate(app, db) manager = Manager(app) def make_shell_context(): return dict(app=app, db=db, User=User, Dictionary=Dictionary) manager.add_command("shell", Shell(make_context=make_shell_context)) manager.add_command("db", MigrateCommand) cov = coverage.coverage(branch=True, include="app/*") @manager.command def test(coverage=False): """ Run the unit tests. """ if coverage: cov.start() import unittest tests = unittest.TestLoader().discover("tests") unittest.TextTestRunner(verbosity=2).run(tests) if coverage: cov.stop() cov.save() print("Coverage Summary:") cov.report() cov_dir = os.path.join(basedir, "tmp/coverage") cov.html_report(directory=cov_dir) print("HTML version: %s/index.html" % cov_dir) cov.erase() if __name__ == "__main__": manager.run() ## Instruction: Add Word model to shell context ## Code After: import os import coverage from flask_script import Manager, Shell from flask_migrate import Migrate, MigrateCommand from config import basedir from app import create_app, db from app.models import User, Dictionary, Word app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") migrate = Migrate(app, db) manager = Manager(app) def make_shell_context(): return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word) manager.add_command("shell", Shell(make_context=make_shell_context)) manager.add_command("db", MigrateCommand) cov = coverage.coverage(branch=True, include="app/*") @manager.command def test(coverage=False): """ Run the unit tests. """ if coverage: cov.start() import unittest tests = unittest.TestLoader().discover("tests") unittest.TextTestRunner(verbosity=2).run(tests) if coverage: cov.stop() cov.save() print("Coverage Summary:") cov.report() cov_dir = os.path.join(basedir, "tmp/coverage") cov.html_report(directory=cov_dir) print("HTML version: %s/index.html" % cov_dir) cov.erase() if __name__ == "__main__": manager.run()
... from config import basedir from app import create_app, db from app.models import User, Dictionary, Word app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default") ... def make_shell_context(): return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word) manager.add_command("shell", Shell(make_context=make_shell_context)) ...
c31c54624d7a46dfd9df96e32d2e07246868aecc
tomviz/python/DefaultITKTransform.py
tomviz/python/DefaultITKTransform.py
def transform_scalars(dataset): """Define this method for Python operators that transform the input array.""" from tomviz import utils import numpy as np import itk # Get the current volume as a numpy array. array = utils.get_array(dataset) # Set up some ITK variables itk_image_type = itk.Image.F3 itk_converter = itk.PyBuffer[itk_image_type] # Read the image into ITK itk_image = itk_converter.GetImageFromArray(array) # ITK filter (I have no idea if this is right) filter = \ itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New() filter.SetInitialNeighborhoodRadius(3) filter.SetMultiplier(3) filter.SetNumberOfIterations(25) filter.SetReplaceValue(255) filter.SetSeed((24,65,37)) filter.SetInput(itk_image) filter.Update() # Get the image back from ITK (result is a numpy image) result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput()) # This is where the transformed data is set, it will display in tomviz. utils.set_array(dataset, result)
import tomviz.operators class DefaultITKTransform(tomviz.operators.CancelableOperator): def transform_scalars(self, dataset): """Define this method for Python operators that transform the input array. This example uses an ITK filter to add 10 to each voxel value.""" # Try imports to make sure we have everything that is needed try: from tomviz import itkutils import itk except Exception as exc: print("Could not import necessary module(s)") raise exc self.progress.value = 0 self.progress.maximum = 100 # Add a try/except around the ITK portion. ITK exceptions are # passed up to the Python layer, so we can at least report what # went wrong with the script, e.g., unsupported image type. try: self.progress.value = 0 self.progress.message = "Converting data to ITK image" # Get the ITK image itk_image = itkutils.convert_vtk_to_itk_image(dataset) itk_input_image_type = type(itk_image) self.progress.value = 30 self.progress.message = "Running filter" # ITK filter filter = itk.AddImageFilter[itk_input_image_type, # Input 1 itk_input_image_type, # Input 2 itk_input_image_type].New() # Output filter.SetInput1(itk_image) filter.SetConstant2(10) itkutils.observe_filter_progress(self, filter, 30, 70) try: filter.Update() except RuntimeError: # Exception thrown when ITK filter is aborted return self.progress.message = "Saving results" itkutils.set_array_from_itk_image(dataset, filter.GetOutput()) self.progress.value = 100 except Exception as exc: print("Problem encountered while running %s" % self.__class__.__name__) raise exc
Change the ITK example to use a simpler ITK filter
Change the ITK example to use a simpler ITK filter
Python
bsd-3-clause
cjh1/tomviz,cryos/tomviz,mathturtle/tomviz,OpenChemistry/tomviz,cjh1/tomviz,thewtex/tomviz,thewtex/tomviz,cryos/tomviz,mathturtle/tomviz,thewtex/tomviz,cjh1/tomviz,cryos/tomviz,OpenChemistry/tomviz,OpenChemistry/tomviz,OpenChemistry/tomviz,mathturtle/tomviz
+ import tomviz.operators - def transform_scalars(dataset): - """Define this method for Python operators that - transform the input array.""" - from tomviz import utils - import numpy as np - import itk - # Get the current volume as a numpy array. - array = utils.get_array(dataset) + class DefaultITKTransform(tomviz.operators.CancelableOperator): - # Set up some ITK variables - itk_image_type = itk.Image.F3 - itk_converter = itk.PyBuffer[itk_image_type] - # Read the image into ITK - itk_image = itk_converter.GetImageFromArray(array) + def transform_scalars(self, dataset): + """Define this method for Python operators that transform the input + array. This example uses an ITK filter to add 10 to each voxel value.""" + # Try imports to make sure we have everything that is needed + try: + from tomviz import itkutils + import itk + except Exception as exc: + print("Could not import necessary module(s)") + raise exc - # ITK filter (I have no idea if this is right) - filter = \ - itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New() - filter.SetInitialNeighborhoodRadius(3) - filter.SetMultiplier(3) - filter.SetNumberOfIterations(25) - filter.SetReplaceValue(255) - filter.SetSeed((24,65,37)) - filter.SetInput(itk_image) - filter.Update() - # Get the image back from ITK (result is a numpy image) - result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput()) + self.progress.value = 0 + self.progress.maximum = 100 - # This is where the transformed data is set, it will display in tomviz. - utils.set_array(dataset, result) + # Add a try/except around the ITK portion. ITK exceptions are + # passed up to the Python layer, so we can at least report what + # went wrong with the script, e.g., unsupported image type. + try: + self.progress.value = 0 + self.progress.message = "Converting data to ITK image" + # Get the ITK image + itk_image = itkutils.convert_vtk_to_itk_image(dataset) + itk_input_image_type = type(itk_image) + self.progress.value = 30 + self.progress.message = "Running filter" + + # ITK filter + filter = itk.AddImageFilter[itk_input_image_type, # Input 1 + itk_input_image_type, # Input 2 + itk_input_image_type].New() # Output + filter.SetInput1(itk_image) + filter.SetConstant2(10) + itkutils.observe_filter_progress(self, filter, 30, 70) + + try: + filter.Update() + except RuntimeError: # Exception thrown when ITK filter is aborted + return + + self.progress.message = "Saving results" + + itkutils.set_array_from_itk_image(dataset, filter.GetOutput()) + + self.progress.value = 100 + except Exception as exc: + print("Problem encountered while running %s" % + self.__class__.__name__) + raise exc +
Change the ITK example to use a simpler ITK filter
## Code Before: def transform_scalars(dataset): """Define this method for Python operators that transform the input array.""" from tomviz import utils import numpy as np import itk # Get the current volume as a numpy array. array = utils.get_array(dataset) # Set up some ITK variables itk_image_type = itk.Image.F3 itk_converter = itk.PyBuffer[itk_image_type] # Read the image into ITK itk_image = itk_converter.GetImageFromArray(array) # ITK filter (I have no idea if this is right) filter = \ itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New() filter.SetInitialNeighborhoodRadius(3) filter.SetMultiplier(3) filter.SetNumberOfIterations(25) filter.SetReplaceValue(255) filter.SetSeed((24,65,37)) filter.SetInput(itk_image) filter.Update() # Get the image back from ITK (result is a numpy image) result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput()) # This is where the transformed data is set, it will display in tomviz. utils.set_array(dataset, result) ## Instruction: Change the ITK example to use a simpler ITK filter ## Code After: import tomviz.operators class DefaultITKTransform(tomviz.operators.CancelableOperator): def transform_scalars(self, dataset): """Define this method for Python operators that transform the input array. This example uses an ITK filter to add 10 to each voxel value.""" # Try imports to make sure we have everything that is needed try: from tomviz import itkutils import itk except Exception as exc: print("Could not import necessary module(s)") raise exc self.progress.value = 0 self.progress.maximum = 100 # Add a try/except around the ITK portion. ITK exceptions are # passed up to the Python layer, so we can at least report what # went wrong with the script, e.g., unsupported image type. try: self.progress.value = 0 self.progress.message = "Converting data to ITK image" # Get the ITK image itk_image = itkutils.convert_vtk_to_itk_image(dataset) itk_input_image_type = type(itk_image) self.progress.value = 30 self.progress.message = "Running filter" # ITK filter filter = itk.AddImageFilter[itk_input_image_type, # Input 1 itk_input_image_type, # Input 2 itk_input_image_type].New() # Output filter.SetInput1(itk_image) filter.SetConstant2(10) itkutils.observe_filter_progress(self, filter, 30, 70) try: filter.Update() except RuntimeError: # Exception thrown when ITK filter is aborted return self.progress.message = "Saving results" itkutils.set_array_from_itk_image(dataset, filter.GetOutput()) self.progress.value = 100 except Exception as exc: print("Problem encountered while running %s" % self.__class__.__name__) raise exc
// ... existing code ... import tomviz.operators class DefaultITKTransform(tomviz.operators.CancelableOperator): def transform_scalars(self, dataset): """Define this method for Python operators that transform the input array. This example uses an ITK filter to add 10 to each voxel value.""" # Try imports to make sure we have everything that is needed try: from tomviz import itkutils import itk except Exception as exc: print("Could not import necessary module(s)") raise exc self.progress.value = 0 self.progress.maximum = 100 # Add a try/except around the ITK portion. ITK exceptions are # passed up to the Python layer, so we can at least report what # went wrong with the script, e.g., unsupported image type. try: self.progress.value = 0 self.progress.message = "Converting data to ITK image" # Get the ITK image itk_image = itkutils.convert_vtk_to_itk_image(dataset) itk_input_image_type = type(itk_image) self.progress.value = 30 self.progress.message = "Running filter" # ITK filter filter = itk.AddImageFilter[itk_input_image_type, # Input 1 itk_input_image_type, # Input 2 itk_input_image_type].New() # Output filter.SetInput1(itk_image) filter.SetConstant2(10) itkutils.observe_filter_progress(self, filter, 30, 70) try: filter.Update() except RuntimeError: # Exception thrown when ITK filter is aborted return self.progress.message = "Saving results" itkutils.set_array_from_itk_image(dataset, filter.GetOutput()) self.progress.value = 100 except Exception as exc: print("Problem encountered while running %s" % self.__class__.__name__) raise exc // ... rest of the code ...
0887e200f31edd8d61e0dd1d3fefae7e828c9269
mindbender/maya/plugins/validate_single_assembly.py
mindbender/maya/plugins/validate_single_assembly.py
import pyblish.api class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin): """Each asset must have a single top-level group The given instance is test-exported, along with construction history to test whether more than 1 top-level DAG node would be included in the exported file. """ label = "Validate Single Assembly" order = pyblish.api.ValidatorOrder hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] def process(self, instance): from maya import cmds from mindbender import maya with maya.maintained_selection(): cmds.select(instance, replace=True) nodes = cmds.file( constructionHistory=True, exportSelected=True, preview=True, force=True, ) assemblies = cmds.ls(nodes, assemblies=True) if not assemblies: raise Exception("No assembly found.") if len(assemblies) != 1: assemblies = '"%s"' % '", "'.join(assemblies) raise Exception( "Multiple assemblies found: %s" % assemblies )
import pyblish.api class SelectAssemblies(pyblish.api.Action): label = "Select Assemblies" on = "failed" def process(self, context, plugin): from maya import cmds cmds.select(plugin.assemblies) class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin): """Each asset must have a single top-level group The given instance is test-exported, along with construction history to test whether more than 1 top-level DAG node would be included in the exported file. """ label = "Validate Single Assembly" order = pyblish.api.ValidatorOrder hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] actions = [ pyblish.api.Category("Actions"), SelectAssemblies, ] assemblies = [] def process(self, instance): from maya import cmds from mindbender import maya with maya.maintained_selection(): cmds.select(instance, replace=True) nodes = cmds.file( constructionHistory=True, exportSelected=True, preview=True, force=True, ) self.assemblies[:] = cmds.ls(nodes, assemblies=True) if not self.assemblies: raise Exception("No assembly found.") if len(self.assemblies) != 1: self.assemblies = '"%s"' % '", "'.join(self.assemblies) raise Exception( "Multiple assemblies found: %s" % self.assemblies )
Add action to select the multiple assemblies.
Add action to select the multiple assemblies.
Python
mit
getavalon/core,MoonShineVFX/core,MoonShineVFX/core,mindbender-studio/core,mindbender-studio/core,getavalon/core
import pyblish.api + + + class SelectAssemblies(pyblish.api.Action): + label = "Select Assemblies" + on = "failed" + + def process(self, context, plugin): + from maya import cmds + cmds.select(plugin.assemblies) class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin): """Each asset must have a single top-level group The given instance is test-exported, along with construction history to test whether more than 1 top-level DAG node would be included in the exported file. """ label = "Validate Single Assembly" order = pyblish.api.ValidatorOrder hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] + actions = [ + pyblish.api.Category("Actions"), + SelectAssemblies, + ] + + assemblies = [] def process(self, instance): from maya import cmds from mindbender import maya with maya.maintained_selection(): cmds.select(instance, replace=True) nodes = cmds.file( constructionHistory=True, exportSelected=True, preview=True, force=True, ) - assemblies = cmds.ls(nodes, assemblies=True) + self.assemblies[:] = cmds.ls(nodes, assemblies=True) - if not assemblies: + if not self.assemblies: raise Exception("No assembly found.") - if len(assemblies) != 1: + if len(self.assemblies) != 1: - assemblies = '"%s"' % '", "'.join(assemblies) + self.assemblies = '"%s"' % '", "'.join(self.assemblies) raise Exception( - "Multiple assemblies found: %s" % assemblies + "Multiple assemblies found: %s" % self.assemblies )
Add action to select the multiple assemblies.
## Code Before: import pyblish.api class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin): """Each asset must have a single top-level group The given instance is test-exported, along with construction history to test whether more than 1 top-level DAG node would be included in the exported file. """ label = "Validate Single Assembly" order = pyblish.api.ValidatorOrder hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] def process(self, instance): from maya import cmds from mindbender import maya with maya.maintained_selection(): cmds.select(instance, replace=True) nodes = cmds.file( constructionHistory=True, exportSelected=True, preview=True, force=True, ) assemblies = cmds.ls(nodes, assemblies=True) if not assemblies: raise Exception("No assembly found.") if len(assemblies) != 1: assemblies = '"%s"' % '", "'.join(assemblies) raise Exception( "Multiple assemblies found: %s" % assemblies ) ## Instruction: Add action to select the multiple assemblies. ## Code After: import pyblish.api class SelectAssemblies(pyblish.api.Action): label = "Select Assemblies" on = "failed" def process(self, context, plugin): from maya import cmds cmds.select(plugin.assemblies) class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin): """Each asset must have a single top-level group The given instance is test-exported, along with construction history to test whether more than 1 top-level DAG node would be included in the exported file. """ label = "Validate Single Assembly" order = pyblish.api.ValidatorOrder hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] actions = [ pyblish.api.Category("Actions"), SelectAssemblies, ] assemblies = [] def process(self, instance): from maya import cmds from mindbender import maya with maya.maintained_selection(): cmds.select(instance, replace=True) nodes = cmds.file( constructionHistory=True, exportSelected=True, preview=True, force=True, ) self.assemblies[:] = cmds.ls(nodes, assemblies=True) if not self.assemblies: raise Exception("No assembly found.") if len(self.assemblies) != 1: self.assemblies = '"%s"' % '", "'.join(self.assemblies) raise Exception( "Multiple assemblies found: %s" % self.assemblies )
# ... existing code ... import pyblish.api class SelectAssemblies(pyblish.api.Action): label = "Select Assemblies" on = "failed" def process(self, context, plugin): from maya import cmds cmds.select(plugin.assemblies) # ... modified code ... hosts = ["maya"] families = ["mindbender.model", "mindbender.rig"] actions = [ pyblish.api.Category("Actions"), SelectAssemblies, ] assemblies = [] def process(self, instance): ... ) self.assemblies[:] = cmds.ls(nodes, assemblies=True) if not self.assemblies: raise Exception("No assembly found.") if len(self.assemblies) != 1: self.assemblies = '"%s"' % '", "'.join(self.assemblies) raise Exception( "Multiple assemblies found: %s" % self.assemblies ) # ... rest of the code ...
b2bc77023ed3e19f6f7483645e2a11952c061de0
tests/registryd/test_registry_startup.py
tests/registryd/test_registry_startup.py
PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties' ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' def test_accessible_iface_properties(registry, session_manager): val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE) assert str(val) == 'main'
PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties' ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' def get_property(proxy, iface_name, prop_name): return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE) def test_accessible_iface_properties(registry, session_manager): values = [ ('Name', 'main'), ('Description', ''), ] for prop_name, expected in values: assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
Test the Description property of the registry's root
Test the Description property of the registry's root
Python
lgpl-2.1
GNOME/at-spi2-core,GNOME/at-spi2-core,GNOME/at-spi2-core
PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties' ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' + def get_property(proxy, iface_name, prop_name): + return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE) + def test_accessible_iface_properties(registry, session_manager): - val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE) - assert str(val) == 'main' + values = [ + ('Name', 'main'), + ('Description', ''), + ] + for prop_name, expected in values: + assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected +
Test the Description property of the registry's root
## Code Before: PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties' ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' def test_accessible_iface_properties(registry, session_manager): val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE) assert str(val) == 'main' ## Instruction: Test the Description property of the registry's root ## Code After: PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties' ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' def get_property(proxy, iface_name, prop_name): return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE) def test_accessible_iface_properties(registry, session_manager): values = [ ('Name', 'main'), ('Description', ''), ] for prop_name, expected in values: assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
# ... existing code ... ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible' def get_property(proxy, iface_name, prop_name): return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE) def test_accessible_iface_properties(registry, session_manager): values = [ ('Name', 'main'), ('Description', ''), ] for prop_name, expected in values: assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected # ... rest of the code ...
805c52698b3fed8df98462c15045f5de3822e241
edx_repo_tools/dev/clone_org.py
edx_repo_tools/dev/clone_org.py
"""Clone an entire GitHub organization.""" import os.path import click from git.repo.base import Repo from edx_repo_tools.auth import pass_github @click.command() @click.option( '--forks/--no-forks', is_flag=True, default=False, help="Should forks be included?" ) @click.option( '--depth', type=int, default=0, help="Depth argument for git clone", ) @click.argument( 'org' ) @pass_github def main(hub, forks, depth, org): for repo in hub.organization(org).repositories(): if repo.fork and not forks: continue dir_name = repo.name dir_name = dir_name.lstrip("-") # avoid dirname/option confusion if os.path.exists(dir_name): continue print(repo.full_name) clone_args = {} if depth: clone_args['depth'] = depth Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
"""Clone an entire GitHub organization.""" import os.path import click from git.repo.base import Repo from edx_repo_tools.auth import pass_github @click.command() @click.option( '--forks/--no-forks', is_flag=True, default=False, help="Should forks be included?" ) @click.option( '--depth', type=int, default=0, help="Depth argument for git clone", ) @click.argument( 'org' ) @pass_github def main(hub, forks, depth, org): for repo in hub.organization(org).iter_repos(): if repo.fork and not forks: continue dir_name = repo.name dir_name = dir_name.lstrip("-") # avoid dirname/option confusion if os.path.exists(dir_name): continue print(repo.full_name) clone_args = {} if depth: clone_args['depth'] = depth Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
Fix to work in python 3.
Fix to work in python 3.
Python
apache-2.0
edx/repo-tools,edx/repo-tools
"""Clone an entire GitHub organization.""" import os.path import click from git.repo.base import Repo from edx_repo_tools.auth import pass_github - @click.command() @click.option( '--forks/--no-forks', is_flag=True, default=False, help="Should forks be included?" ) @click.option( '--depth', type=int, default=0, help="Depth argument for git clone", ) @click.argument( 'org' ) @pass_github def main(hub, forks, depth, org): - for repo in hub.organization(org).repositories(): + for repo in hub.organization(org).iter_repos(): if repo.fork and not forks: continue dir_name = repo.name dir_name = dir_name.lstrip("-") # avoid dirname/option confusion if os.path.exists(dir_name): continue print(repo.full_name) clone_args = {} if depth: clone_args['depth'] = depth Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
Fix to work in python 3.
## Code Before: """Clone an entire GitHub organization.""" import os.path import click from git.repo.base import Repo from edx_repo_tools.auth import pass_github @click.command() @click.option( '--forks/--no-forks', is_flag=True, default=False, help="Should forks be included?" ) @click.option( '--depth', type=int, default=0, help="Depth argument for git clone", ) @click.argument( 'org' ) @pass_github def main(hub, forks, depth, org): for repo in hub.organization(org).repositories(): if repo.fork and not forks: continue dir_name = repo.name dir_name = dir_name.lstrip("-") # avoid dirname/option confusion if os.path.exists(dir_name): continue print(repo.full_name) clone_args = {} if depth: clone_args['depth'] = depth Repo.clone_from(repo.ssh_url, dir_name, **clone_args) ## Instruction: Fix to work in python 3. ## Code After: """Clone an entire GitHub organization.""" import os.path import click from git.repo.base import Repo from edx_repo_tools.auth import pass_github @click.command() @click.option( '--forks/--no-forks', is_flag=True, default=False, help="Should forks be included?" ) @click.option( '--depth', type=int, default=0, help="Depth argument for git clone", ) @click.argument( 'org' ) @pass_github def main(hub, forks, depth, org): for repo in hub.organization(org).iter_repos(): if repo.fork and not forks: continue dir_name = repo.name dir_name = dir_name.lstrip("-") # avoid dirname/option confusion if os.path.exists(dir_name): continue print(repo.full_name) clone_args = {} if depth: clone_args['depth'] = depth Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
// ... existing code ... from edx_repo_tools.auth import pass_github @click.command() // ... modified code ... @pass_github def main(hub, forks, depth, org): for repo in hub.organization(org).iter_repos(): if repo.fork and not forks: continue // ... rest of the code ...
66946f72d243f1836df0dbd8917f204011ec1701
hs_core/autocomplete_light_registry.py
hs_core/autocomplete_light_registry.py
from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group class UserAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['username', 'first_name', 'last_name'] split_words = True def choices_for_request(self): self.choices = self.choices.filter(is_active=True) return super(UserAutocomplete, self).choices_for_request() def choice_label(self, choice): label = "" if choice.first_name: label += choice.first_name if choice.last_name: if choice.first_name: label += " " label += choice.last_name if choice.userprofile.organization: if choice.first_name or choice.last_name: label += ", " label += choice.userprofile.organization if choice.username: label += "".join([" (", choice.username, ")"]) return label autocomplete_light.register(User, UserAutocomplete) class GroupAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields=['name'] def choices_for_request(self): self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author') return super(GroupAutocomplete, self).choices_for_request() autocomplete_light.register(Group, GroupAutocomplete)
from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group class UserAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['username', 'first_name', 'last_name'] split_words = True def choices_for_request(self): self.choices = self.choices.filter(is_active=True) return super(UserAutocomplete, self).choices_for_request() def choice_label(self, choice): label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""]) if choice.userprofile.organization: if choice.first_name or choice.last_name: label += ", " label += choice.userprofile.organization if choice.username: label += "".join([" (", choice.username, ")"]) return label autocomplete_light.register(User, UserAutocomplete) class GroupAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields=['name'] def choices_for_request(self): self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author') return super(GroupAutocomplete, self).choices_for_request() autocomplete_light.register(Group, GroupAutocomplete)
Add middle name display to autocomplete widget
Add middle name display to autocomplete widget
Python
bsd-3-clause
hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare
from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group + class UserAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['username', 'first_name', 'last_name'] split_words = True def choices_for_request(self): self.choices = self.choices.filter(is_active=True) return super(UserAutocomplete, self).choices_for_request() def choice_label(self, choice): + label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""]) - label = "" - - if choice.first_name: - label += choice.first_name - - if choice.last_name: - if choice.first_name: - label += " " - label += choice.last_name if choice.userprofile.organization: if choice.first_name or choice.last_name: label += ", " label += choice.userprofile.organization if choice.username: label += "".join([" (", choice.username, ")"]) return label autocomplete_light.register(User, UserAutocomplete) class GroupAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields=['name'] def choices_for_request(self): self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author') return super(GroupAutocomplete, self).choices_for_request() autocomplete_light.register(Group, GroupAutocomplete)
Add middle name display to autocomplete widget
## Code Before: from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group class UserAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['username', 'first_name', 'last_name'] split_words = True def choices_for_request(self): self.choices = self.choices.filter(is_active=True) return super(UserAutocomplete, self).choices_for_request() def choice_label(self, choice): label = "" if choice.first_name: label += choice.first_name if choice.last_name: if choice.first_name: label += " " label += choice.last_name if choice.userprofile.organization: if choice.first_name or choice.last_name: label += ", " label += choice.userprofile.organization if choice.username: label += "".join([" (", choice.username, ")"]) return label autocomplete_light.register(User, UserAutocomplete) class GroupAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields=['name'] def choices_for_request(self): self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author') return super(GroupAutocomplete, self).choices_for_request() autocomplete_light.register(Group, GroupAutocomplete) ## Instruction: Add middle name display to autocomplete widget ## Code After: from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group class UserAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['username', 'first_name', 'last_name'] split_words = True def choices_for_request(self): self.choices = self.choices.filter(is_active=True) return super(UserAutocomplete, self).choices_for_request() def choice_label(self, choice): label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""]) if choice.userprofile.organization: if choice.first_name or choice.last_name: label += ", " label += choice.userprofile.organization if choice.username: label += "".join([" (", choice.username, ")"]) return label autocomplete_light.register(User, UserAutocomplete) class GroupAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields=['name'] def choices_for_request(self): self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author') return super(GroupAutocomplete, self).choices_for_request() autocomplete_light.register(Group, GroupAutocomplete)
... from autocomplete_light import shortcuts as autocomplete_light from django.contrib.auth.models import User, Group class UserAutocomplete(autocomplete_light.AutocompleteModelBase): ... def choice_label(self, choice): label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""]) if choice.userprofile.organization: ...
bddab649c6684f09870983dca97c39eb30b62c06
djangobotcfg/status.py
djangobotcfg/status.py
from buildbot.status import html, words from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth # authz = Authz( # forceBuild=True, # forceAllBuilds=True, # pingBuilder=True, # gracefulShutdown=True, # stopBuild=True, # stopAllBuilds=True, # cancelPendingBuild=True, # cleanShutdown=True, # ) def get_status(): return [ html.WebStatus( http_port = '8010', # authz = authz, order_console_by_time = True, revlink = 'http://code.djangoproject.com/changeset/%s', changecommentlink = ( r'\b#(\d+)\b', r'http://code.djangoproject.com/ticket/\1', r'Ticket \g<0>' ) ), words.IRC( host = 'irc.freenode.net', channels = ['#revsys'], nick = 'djangobuilds', notify_events = { 'successToFailure': True, 'failureToSuccess': True, } ) ]
from buildbot.status import html, words from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth def get_status(): return [ html.WebStatus( http_port = '8010', # authz = authz, order_console_by_time = True, revlink = 'http://code.djangoproject.com/changeset/%s', changecommentlink = ( r'\b#(\d+)\b', r'http://code.djangoproject.com/ticket/\1', r'Ticket \g<0>' ) ), ]
Remove the IRC bot for now, and also the commented-out code.
Remove the IRC bot for now, and also the commented-out code.
Python
bsd-3-clause
hochanh/django-buildmaster,jacobian-archive/django-buildmaster
from buildbot.status import html, words from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth - - # authz = Authz( - # forceBuild=True, - # forceAllBuilds=True, - # pingBuilder=True, - # gracefulShutdown=True, - # stopBuild=True, - # stopAllBuilds=True, - # cancelPendingBuild=True, - # cleanShutdown=True, - # ) def get_status(): return [ html.WebStatus( http_port = '8010', # authz = authz, order_console_by_time = True, revlink = 'http://code.djangoproject.com/changeset/%s', changecommentlink = ( r'\b#(\d+)\b', r'http://code.djangoproject.com/ticket/\1', r'Ticket \g<0>' ) ), - - words.IRC( - host = 'irc.freenode.net', - channels = ['#revsys'], - nick = 'djangobuilds', - notify_events = { - 'successToFailure': True, - 'failureToSuccess': True, - } - ) ]
Remove the IRC bot for now, and also the commented-out code.
## Code Before: from buildbot.status import html, words from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth # authz = Authz( # forceBuild=True, # forceAllBuilds=True, # pingBuilder=True, # gracefulShutdown=True, # stopBuild=True, # stopAllBuilds=True, # cancelPendingBuild=True, # cleanShutdown=True, # ) def get_status(): return [ html.WebStatus( http_port = '8010', # authz = authz, order_console_by_time = True, revlink = 'http://code.djangoproject.com/changeset/%s', changecommentlink = ( r'\b#(\d+)\b', r'http://code.djangoproject.com/ticket/\1', r'Ticket \g<0>' ) ), words.IRC( host = 'irc.freenode.net', channels = ['#revsys'], nick = 'djangobuilds', notify_events = { 'successToFailure': True, 'failureToSuccess': True, } ) ] ## Instruction: Remove the IRC bot for now, and also the commented-out code. ## Code After: from buildbot.status import html, words from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth def get_status(): return [ html.WebStatus( http_port = '8010', # authz = authz, order_console_by_time = True, revlink = 'http://code.djangoproject.com/changeset/%s', changecommentlink = ( r'\b#(\d+)\b', r'http://code.djangoproject.com/ticket/\1', r'Ticket \g<0>' ) ), ]
... from buildbot.status.web.authz import Authz from buildbot.status.web.auth import BasicAuth def get_status(): ... ) ), ] ...
a4375a6ec5ca54b887527885235317986011801c
guesser.py
guesser.py
from synt.utils.redis_manager import RedisManager from synt.utils.extractors import best_word_feats from synt.utils.text import sanitize_text MANAGER = RedisManager() DEFAULT_CLASSIFIER = MANAGER.load_classifier() def guess(text, classifier=DEFAULT_CLASSIFIER): """Takes a blob of text and returns the sentiment and confidence score.""" assert classifier, "Needs a classifier." bag_of_words = best_word_feats(sanitize_text(text)) if bag_of_words: guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) return (guess, [(prob.prob(sample),sample) for sample in prob.samples()])
from synt.utils.redis_manager import RedisManager from synt.utils.extractors import best_word_feats from synt.utils.text import sanitize_text MANAGER = RedisManager() DEFAULT_CLASSIFIER = MANAGER.load_classifier() def guess(text, classifier=DEFAULT_CLASSIFIER): """Takes a blob of text and returns the sentiment and confidence score.""" assert classifier, "Needs a classifier." bag_of_words = best_word_feats(sanitize_text(text)) if bag_of_words: guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) #return a -1 .. 1 score score = prob.prob('positive') - prob.prob('negative') return score
Return a -1 .. 1 sentiment score.
Return a -1 .. 1 sentiment score.
Python
agpl-3.0
lrvick/synt
from synt.utils.redis_manager import RedisManager from synt.utils.extractors import best_word_feats from synt.utils.text import sanitize_text MANAGER = RedisManager() DEFAULT_CLASSIFIER = MANAGER.load_classifier() def guess(text, classifier=DEFAULT_CLASSIFIER): """Takes a blob of text and returns the sentiment and confidence score.""" assert classifier, "Needs a classifier." bag_of_words = best_word_feats(sanitize_text(text)) if bag_of_words: guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) - return (guess, [(prob.prob(sample),sample) for sample in prob.samples()]) + + #return a -1 .. 1 score + score = prob.prob('positive') - prob.prob('negative') + + return score
Return a -1 .. 1 sentiment score.
## Code Before: from synt.utils.redis_manager import RedisManager from synt.utils.extractors import best_word_feats from synt.utils.text import sanitize_text MANAGER = RedisManager() DEFAULT_CLASSIFIER = MANAGER.load_classifier() def guess(text, classifier=DEFAULT_CLASSIFIER): """Takes a blob of text and returns the sentiment and confidence score.""" assert classifier, "Needs a classifier." bag_of_words = best_word_feats(sanitize_text(text)) if bag_of_words: guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) return (guess, [(prob.prob(sample),sample) for sample in prob.samples()]) ## Instruction: Return a -1 .. 1 sentiment score. ## Code After: from synt.utils.redis_manager import RedisManager from synt.utils.extractors import best_word_feats from synt.utils.text import sanitize_text MANAGER = RedisManager() DEFAULT_CLASSIFIER = MANAGER.load_classifier() def guess(text, classifier=DEFAULT_CLASSIFIER): """Takes a blob of text and returns the sentiment and confidence score.""" assert classifier, "Needs a classifier." bag_of_words = best_word_feats(sanitize_text(text)) if bag_of_words: guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) #return a -1 .. 1 score score = prob.prob('positive') - prob.prob('negative') return score
# ... existing code ... guess = classifier.classify(bag_of_words) prob = classifier.prob_classify(bag_of_words) #return a -1 .. 1 score score = prob.prob('positive') - prob.prob('negative') return score # ... rest of the code ...
f8bdd7c8139cfc6d7af4bb3d89e983073db976bf
mecodesktop.py
mecodesktop.py
from macroeco import desktop desktop()
import sys as _sys from macroeco import desktop if len(_sys.argv) > 1: desktop(_sys.argv[1]) else: desktop()
Allow compiled OS X app to take parameter file as input on command line
Allow compiled OS X app to take parameter file as input on command line
Python
bsd-2-clause
jkitzes/macroeco
+ import sys as _sys from macroeco import desktop - desktop() + if len(_sys.argv) > 1: + desktop(_sys.argv[1]) + else: + desktop() +
Allow compiled OS X app to take parameter file as input on command line
## Code Before: from macroeco import desktop desktop() ## Instruction: Allow compiled OS X app to take parameter file as input on command line ## Code After: import sys as _sys from macroeco import desktop if len(_sys.argv) > 1: desktop(_sys.argv[1]) else: desktop()
... import sys as _sys from macroeco import desktop if len(_sys.argv) > 1: desktop(_sys.argv[1]) else: desktop() ...
f3ea9820a96536e74e6f74f13387140c97ea9f2e
backgroundworker.py
backgroundworker.py
import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = engine, if_exists = 'replace')
import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = connection, if_exists = 'replace') connection.close()
Add engine connection, and close engine connection to worker dyno
Add engine connection, and close engine connection to worker dyno
Python
mit
caseymacphee/green_quote,caseymacphee/green_quote
import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) + engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') + connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] - df.to_sql(name='entries', con = engine, if_exists = 'replace') + df.to_sql(name='entries', con = connection, if_exists = 'replace') + connection.close() +
Add engine connection, and close engine connection to worker dyno
## Code Before: import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = engine, if_exists = 'replace') ## Instruction: Add engine connection, and close engine connection to worker dyno ## Code After: import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = connection, if_exists = 'replace') connection.close()
// ... existing code ... # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = connection, if_exists = 'replace') connection.close() // ... rest of the code ...
24fc06d17303868ef4ea057cd001ec6cb49ab18f
flask_app.py
flask_app.py
import os from flask import Flask, render_template from jinja2 import Template app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..') app.config.from_pyfile('settings.py') BASE = '/%s' % app.config['REPO_NAME'] @app.route('/') def home(): with open('talk.md', 'r') as f: template = Template(f.read()) markdown = template.render(base=BASE) js_file = 'talk.js' if os.path.isfile(js_file): with open(js_file, 'r') as f_js: js = f_js.read() else: js = '' return render_template('slides.html', markdown=markdown, js=js) if __name__ == '__main__': BASE = '' port = int(os.environ.get('PORT', 5000)) app.run(host='0.0.0.0', port=port)
import os from flask import Flask, render_template from jinja2 import Template app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..') app.config.from_pyfile('settings.py') BASE = '/%s' % app.config['REPO_NAME'] @app.route('/') def home(): with open('talk.md', 'r') as f: template = Template(f.read().decode('utf-8')) markdown = template.render(base=BASE) js_file = 'talk.js' if os.path.isfile(js_file): with open(js_file, 'r') as f_js: js = f_js.read() else: js = '' return render_template('slides.html', markdown=markdown, js=js) if __name__ == '__main__': BASE = '' port = int(os.environ.get('PORT', 5000)) app.run(host='0.0.0.0', port=port)
Fix utf-8 problem with åäö and friends.
Fix utf-8 problem with åäö and friends.
Python
bsd-3-clause
sknippen/refreeze,sknippen/refreeze,sknippen/refreeze
import os from flask import Flask, render_template from jinja2 import Template app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..') app.config.from_pyfile('settings.py') BASE = '/%s' % app.config['REPO_NAME'] @app.route('/') def home(): with open('talk.md', 'r') as f: - template = Template(f.read()) + template = Template(f.read().decode('utf-8')) markdown = template.render(base=BASE) js_file = 'talk.js' if os.path.isfile(js_file): with open(js_file, 'r') as f_js: js = f_js.read() else: js = '' return render_template('slides.html', markdown=markdown, js=js) if __name__ == '__main__': BASE = '' port = int(os.environ.get('PORT', 5000)) app.run(host='0.0.0.0', port=port)
Fix utf-8 problem with åäö and friends.
## Code Before: import os from flask import Flask, render_template from jinja2 import Template app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..') app.config.from_pyfile('settings.py') BASE = '/%s' % app.config['REPO_NAME'] @app.route('/') def home(): with open('talk.md', 'r') as f: template = Template(f.read()) markdown = template.render(base=BASE) js_file = 'talk.js' if os.path.isfile(js_file): with open(js_file, 'r') as f_js: js = f_js.read() else: js = '' return render_template('slides.html', markdown=markdown, js=js) if __name__ == '__main__': BASE = '' port = int(os.environ.get('PORT', 5000)) app.run(host='0.0.0.0', port=port) ## Instruction: Fix utf-8 problem with åäö and friends. ## Code After: import os from flask import Flask, render_template from jinja2 import Template app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..') app.config.from_pyfile('settings.py') BASE = '/%s' % app.config['REPO_NAME'] @app.route('/') def home(): with open('talk.md', 'r') as f: template = Template(f.read().decode('utf-8')) markdown = template.render(base=BASE) js_file = 'talk.js' if os.path.isfile(js_file): with open(js_file, 'r') as f_js: js = f_js.read() else: js = '' return render_template('slides.html', markdown=markdown, js=js) if __name__ == '__main__': BASE = '' port = int(os.environ.get('PORT', 5000)) app.run(host='0.0.0.0', port=port)
// ... existing code ... def home(): with open('talk.md', 'r') as f: template = Template(f.read().decode('utf-8')) markdown = template.render(base=BASE) js_file = 'talk.js' // ... rest of the code ...
41b241de6f2afa94b442007518d481526bfb66ae
linked-list/remove-k-from-list.py
linked-list/remove-k-from-list.py
class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None
class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
Add initialization to linked list class
Add initialization to linked list class
Python
mit
derekmpham/interview-prep,derekmpham/interview-prep
class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None + class LinkedList(object): + def __init__(self, head=None): + self.head = head +
Add initialization to linked list class
## Code Before: class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None ## Instruction: Add initialization to linked list class ## Code After: class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
# ... existing code ... self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head # ... rest of the code ...
697c590bf60c261280e55f8580b33423dbe800c6
splinter/driver/webdriver/firefox.py
splinter/driver/webdriver/firefox.py
import subprocess from selenium.webdriver import Firefox from selenium.webdriver.firefox.firefox_profile import FirefoxProfile from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement from splinter.driver.webdriver.cookie_manager import CookieManager class WebDriver(BaseWebDriver): def __init__(self, profile=None, extensions=None): self.old_popen = subprocess.Popen firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) if extensions: for extension in extensions: firefox_profile.add_extension(extension) self._patch_subprocess() self.driver = Firefox(firefox_profile) self._unpatch_subprocess() self.element_class = WebDriverElement self._cookie_manager = CookieManager(self.driver) super(WebDriver, self).__init__() class WebDriverElement(BaseWebDriverElement): def mouseover(self): """ Firefox doesn't support mouseover. """ raise NotImplementedError("Firefox doesn't support mouse over") def mouseout(self): """ Firefox doesn't support mouseout. """ raise NotImplementedError("Firefox doesn't support mouseout") def double_click(self): """ Firefox doesn't support doubleclick. """ raise NotImplementedError("Firefox doesn't support doubleclick")
import subprocess from selenium.webdriver import Firefox from selenium.webdriver.firefox.firefox_profile import FirefoxProfile from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement from splinter.driver.webdriver.cookie_manager import CookieManager class WebDriver(BaseWebDriver): def __init__(self, profile=None, extensions=None): self.old_popen = subprocess.Popen firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) firefox_profile.set_preference('network.dns.disableIPv6', False) if extensions: for extension in extensions: firefox_profile.add_extension(extension) self._patch_subprocess() self.driver = Firefox(firefox_profile) self._unpatch_subprocess() self.element_class = WebDriverElement self._cookie_manager = CookieManager(self.driver) super(WebDriver, self).__init__() class WebDriverElement(BaseWebDriverElement): def mouseover(self): """ Firefox doesn't support mouseover. """ raise NotImplementedError("Firefox doesn't support mouse over") def mouseout(self): """ Firefox doesn't support mouseout. """ raise NotImplementedError("Firefox doesn't support mouseout") def double_click(self): """ Firefox doesn't support doubleclick. """ raise NotImplementedError("Firefox doesn't support doubleclick")
Fix error on Firefox 6 where pages are not open if this preference is True (default).
Fix error on Firefox 6 where pages are not open if this preference is True (default).
Python
bsd-3-clause
bmcculley/splinter,cobrateam/splinter,bmcculley/splinter,nikolas/splinter,drptbl/splinter,objarni/splinter,nikolas/splinter,cobrateam/splinter,drptbl/splinter,underdogio/splinter,underdogio/splinter,bubenkoff/splinter,lrowe/splinter,bubenkoff/splinter,lrowe/splinter,objarni/splinter,gjvis/splinter,bmcculley/splinter,underdogio/splinter,objarni/splinter,lrowe/splinter,nikolas/splinter,cobrateam/splinter,gjvis/splinter,myself659/splinter,myself659/splinter,myself659/splinter,gjvis/splinter,drptbl/splinter
import subprocess from selenium.webdriver import Firefox from selenium.webdriver.firefox.firefox_profile import FirefoxProfile from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement from splinter.driver.webdriver.cookie_manager import CookieManager class WebDriver(BaseWebDriver): def __init__(self, profile=None, extensions=None): self.old_popen = subprocess.Popen firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) + firefox_profile.set_preference('network.dns.disableIPv6', False) if extensions: for extension in extensions: firefox_profile.add_extension(extension) self._patch_subprocess() self.driver = Firefox(firefox_profile) self._unpatch_subprocess() self.element_class = WebDriverElement self._cookie_manager = CookieManager(self.driver) super(WebDriver, self).__init__() class WebDriverElement(BaseWebDriverElement): def mouseover(self): """ Firefox doesn't support mouseover. """ raise NotImplementedError("Firefox doesn't support mouse over") def mouseout(self): """ Firefox doesn't support mouseout. """ raise NotImplementedError("Firefox doesn't support mouseout") def double_click(self): """ Firefox doesn't support doubleclick. """ raise NotImplementedError("Firefox doesn't support doubleclick")
Fix error on Firefox 6 where pages are not open if this preference is True (default).
## Code Before: import subprocess from selenium.webdriver import Firefox from selenium.webdriver.firefox.firefox_profile import FirefoxProfile from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement from splinter.driver.webdriver.cookie_manager import CookieManager class WebDriver(BaseWebDriver): def __init__(self, profile=None, extensions=None): self.old_popen = subprocess.Popen firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) if extensions: for extension in extensions: firefox_profile.add_extension(extension) self._patch_subprocess() self.driver = Firefox(firefox_profile) self._unpatch_subprocess() self.element_class = WebDriverElement self._cookie_manager = CookieManager(self.driver) super(WebDriver, self).__init__() class WebDriverElement(BaseWebDriverElement): def mouseover(self): """ Firefox doesn't support mouseover. """ raise NotImplementedError("Firefox doesn't support mouse over") def mouseout(self): """ Firefox doesn't support mouseout. """ raise NotImplementedError("Firefox doesn't support mouseout") def double_click(self): """ Firefox doesn't support doubleclick. """ raise NotImplementedError("Firefox doesn't support doubleclick") ## Instruction: Fix error on Firefox 6 where pages are not open if this preference is True (default). ## Code After: import subprocess from selenium.webdriver import Firefox from selenium.webdriver.firefox.firefox_profile import FirefoxProfile from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement from splinter.driver.webdriver.cookie_manager import CookieManager class WebDriver(BaseWebDriver): def __init__(self, profile=None, extensions=None): self.old_popen = subprocess.Popen firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) firefox_profile.set_preference('network.dns.disableIPv6', False) if extensions: for extension in extensions: firefox_profile.add_extension(extension) self._patch_subprocess() self.driver = Firefox(firefox_profile) self._unpatch_subprocess() self.element_class = WebDriverElement self._cookie_manager = CookieManager(self.driver) super(WebDriver, self).__init__() class WebDriverElement(BaseWebDriverElement): def mouseover(self): """ Firefox doesn't support mouseover. """ raise NotImplementedError("Firefox doesn't support mouse over") def mouseout(self): """ Firefox doesn't support mouseout. """ raise NotImplementedError("Firefox doesn't support mouseout") def double_click(self): """ Firefox doesn't support doubleclick. """ raise NotImplementedError("Firefox doesn't support doubleclick")
# ... existing code ... firefox_profile = FirefoxProfile(profile) firefox_profile.set_preference('extensions.logging.enabled', False) firefox_profile.set_preference('network.dns.disableIPv6', False) if extensions: # ... rest of the code ...
ea8cbcaf41f01a46390882fbc99e6e14d70a49d1
src/mmw/apps/user/models.py
src/mmw/apps/user/models.py
from django.contrib.auth.models import User from django.db import models class ItsiUserManager(models.Manager): def create_itsi_user(self, user, itsi_id): itsi_user = self.create(user=user, itsi_id=itsi_id) return itsi_user class ItsiUser(models.Model): user = models.OneToOneField(User, primary_key=True) itsi_id = models.IntegerField() objects = ItsiUserManager() def __unicode__(self): return unicode(self.user.username)
from django.conf import settings from django.contrib.auth.models import User from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from rest_framework.authtoken.models import Token @receiver(post_save, sender=settings.AUTH_USER_MODEL) def create_auth_token(sender, instance=None, created=False, **kwargs): """ Create an auth token for every newly created user. """ if created: Token.objects.create(user=instance) class ItsiUserManager(models.Manager): def create_itsi_user(self, user, itsi_id): itsi_user = self.create(user=user, itsi_id=itsi_id) return itsi_user class ItsiUser(models.Model): user = models.OneToOneField(User, primary_key=True) itsi_id = models.IntegerField() objects = ItsiUserManager() def __unicode__(self): return unicode(self.user.username)
Create an API auth token for every newly created user
Create an API auth token for every newly created user * Add a post_save signal to add a new authtoken for every new user. For use with the Geoprocessing API
Python
apache-2.0
WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed
+ + from django.conf import settings from django.contrib.auth.models import User from django.db import models + from django.db.models.signals import post_save + from django.dispatch import receiver + + from rest_framework.authtoken.models import Token + + + @receiver(post_save, sender=settings.AUTH_USER_MODEL) + def create_auth_token(sender, instance=None, created=False, **kwargs): + """ + Create an auth token for every newly created user. + """ + if created: + Token.objects.create(user=instance) class ItsiUserManager(models.Manager): def create_itsi_user(self, user, itsi_id): itsi_user = self.create(user=user, itsi_id=itsi_id) return itsi_user class ItsiUser(models.Model): user = models.OneToOneField(User, primary_key=True) itsi_id = models.IntegerField() objects = ItsiUserManager() def __unicode__(self): return unicode(self.user.username)
Create an API auth token for every newly created user
## Code Before: from django.contrib.auth.models import User from django.db import models class ItsiUserManager(models.Manager): def create_itsi_user(self, user, itsi_id): itsi_user = self.create(user=user, itsi_id=itsi_id) return itsi_user class ItsiUser(models.Model): user = models.OneToOneField(User, primary_key=True) itsi_id = models.IntegerField() objects = ItsiUserManager() def __unicode__(self): return unicode(self.user.username) ## Instruction: Create an API auth token for every newly created user ## Code After: from django.conf import settings from django.contrib.auth.models import User from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from rest_framework.authtoken.models import Token @receiver(post_save, sender=settings.AUTH_USER_MODEL) def create_auth_token(sender, instance=None, created=False, **kwargs): """ Create an auth token for every newly created user. """ if created: Token.objects.create(user=instance) class ItsiUserManager(models.Manager): def create_itsi_user(self, user, itsi_id): itsi_user = self.create(user=user, itsi_id=itsi_id) return itsi_user class ItsiUser(models.Model): user = models.OneToOneField(User, primary_key=True) itsi_id = models.IntegerField() objects = ItsiUserManager() def __unicode__(self): return unicode(self.user.username)
# ... existing code ... from django.conf import settings from django.contrib.auth.models import User from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from rest_framework.authtoken.models import Token @receiver(post_save, sender=settings.AUTH_USER_MODEL) def create_auth_token(sender, instance=None, created=False, **kwargs): """ Create an auth token for every newly created user. """ if created: Token.objects.create(user=instance) # ... rest of the code ...
424980a48e451d1b99397843001bd75fa58e474e
tests/test_fullqualname.py
tests/test_fullqualname.py
"""Tests for fullqualname.""" import nose import sys from fullqualname import fullqualname def test_builtin_function(): # Test built-in function object. obj = len # Type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a function. assert 'built-in function' in repr(obj) if sys.version_info >= (3, ): expected = 'builtins.len' else: expected = '__builtin__.len' nose.tools.assert_equals(fullqualname(obj), expected)
"""Tests for fullqualname.""" import inspect import nose import sys from fullqualname import fullqualname def test_builtin_function(): # Test built-in function object. obj = len # Type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a function. assert 'built-in function' in repr(obj) if sys.version_info >= (3, ): expected = 'builtins.len' else: expected = '__builtin__.len' nose.tools.assert_equals(fullqualname(obj), expected) def test_builtin_method(): # Test built-in method object. obj = [1, 2, 3].append # Object type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a method. assert 'built-in method' in repr(obj) # Object __self__ attribute is not a class. assert not inspect.isclass(obj.__self__) if sys.version_info >= (3, ): expected = 'builtins.list.append' else: expected = '__builtin__.list.append' nose.tools.assert_equals(fullqualname(obj), expected)
Add built-in method object test
Add built-in method object test
Python
bsd-3-clause
etgalloway/fullqualname
"""Tests for fullqualname.""" + import inspect import nose import sys from fullqualname import fullqualname def test_builtin_function(): # Test built-in function object. obj = len # Type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a function. assert 'built-in function' in repr(obj) if sys.version_info >= (3, ): expected = 'builtins.len' else: expected = '__builtin__.len' nose.tools.assert_equals(fullqualname(obj), expected) + + def test_builtin_method(): + # Test built-in method object. + + obj = [1, 2, 3].append + + # Object type is 'builtin_function_or_method'. + assert type(obj).__name__ == 'builtin_function_or_method' + + # Object is a method. + assert 'built-in method' in repr(obj) + + # Object __self__ attribute is not a class. + assert not inspect.isclass(obj.__self__) + + if sys.version_info >= (3, ): + expected = 'builtins.list.append' + else: + expected = '__builtin__.list.append' + + nose.tools.assert_equals(fullqualname(obj), expected) +
Add built-in method object test
## Code Before: """Tests for fullqualname.""" import nose import sys from fullqualname import fullqualname def test_builtin_function(): # Test built-in function object. obj = len # Type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a function. assert 'built-in function' in repr(obj) if sys.version_info >= (3, ): expected = 'builtins.len' else: expected = '__builtin__.len' nose.tools.assert_equals(fullqualname(obj), expected) ## Instruction: Add built-in method object test ## Code After: """Tests for fullqualname.""" import inspect import nose import sys from fullqualname import fullqualname def test_builtin_function(): # Test built-in function object. obj = len # Type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a function. assert 'built-in function' in repr(obj) if sys.version_info >= (3, ): expected = 'builtins.len' else: expected = '__builtin__.len' nose.tools.assert_equals(fullqualname(obj), expected) def test_builtin_method(): # Test built-in method object. obj = [1, 2, 3].append # Object type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a method. assert 'built-in method' in repr(obj) # Object __self__ attribute is not a class. assert not inspect.isclass(obj.__self__) if sys.version_info >= (3, ): expected = 'builtins.list.append' else: expected = '__builtin__.list.append' nose.tools.assert_equals(fullqualname(obj), expected)
# ... existing code ... """Tests for fullqualname.""" import inspect import nose import sys # ... modified code ... nose.tools.assert_equals(fullqualname(obj), expected) def test_builtin_method(): # Test built-in method object. obj = [1, 2, 3].append # Object type is 'builtin_function_or_method'. assert type(obj).__name__ == 'builtin_function_or_method' # Object is a method. assert 'built-in method' in repr(obj) # Object __self__ attribute is not a class. assert not inspect.isclass(obj.__self__) if sys.version_info >= (3, ): expected = 'builtins.list.append' else: expected = '__builtin__.list.append' nose.tools.assert_equals(fullqualname(obj), expected) # ... rest of the code ...
85e8ddb6d72b7f21b49236ea4084029dec09a6f9
projects/forms.py
projects/forms.py
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project fileds = ( 'status', 'attitude', )
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
Exclude fields from the RestrcitedForm (no verification)
Exclude fields from the RestrcitedForm (no verification)
Python
mit
Hackfmi/Diaphanum,Hackfmi/Diaphanum
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project + exclude = ( + 'name', 'team', 'description', 'targets', 'tasks', 'target_group', + 'schedule', 'resources', 'finance_description', 'partners', + 'flp', 'created_at', 'user', + ) + fileds = ( 'status', 'attitude', ) +
Exclude fields from the RestrcitedForm (no verification)
## Code Before: from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project fileds = ( 'status', 'attitude', ) ## Instruction: Exclude fields from the RestrcitedForm (no verification) ## Code After: from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
# ... existing code ... class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', # ... rest of the code ...