commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
3.52k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 7
43k
| ndiff
stringlengths 52
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
| fuzzy_diff
stringlengths 17
3.24k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ff800f11b948808e4574ec3a893ed4e259707dcf | stubs/python2-urllib2/run.py | stubs/python2-urllib2/run.py | import sys
import ssl
import urllib2
if len(sys.argv) < 3 or len(sys.argv) > 4:
exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0])
host = sys.argv[1]
port = sys.argv[2]
cafile = sys.argv[3] if len(sys.argv) > 3 else None
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
except ssl.CertificateError:
print("REJECT")
except urllib2.URLError as exc:
if not isinstance(exc.reason, ssl.SSLError):
raise
print("REJECT")
else:
print("ACCEPT")
| import sys
import ssl
import urllib2
if len(sys.argv) < 3 or len(sys.argv) > 4:
exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0])
host = sys.argv[1]
port = sys.argv[2]
cafile = sys.argv[3] if len(sys.argv) > 3 else None
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
except getattr(ssl, "CertificateError", ()):
print("REJECT")
except urllib2.URLError as exc:
if not isinstance(exc.reason, ssl.SSLError):
raise
print("REJECT")
else:
print("ACCEPT")
| Make python2-urllib2 compatible with more Python 2.7 versions | Make python2-urllib2 compatible with more Python 2.7 versions
Try to catch ssl.CertificateError only if CertificateError is
defined. Otherwise bail out by effectively doing a dummy
"catch ():".
| Python | mit | ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls,ouspg/trytls | import sys
import ssl
import urllib2
if len(sys.argv) < 3 or len(sys.argv) > 4:
exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0])
host = sys.argv[1]
port = sys.argv[2]
cafile = sys.argv[3] if len(sys.argv) > 3 else None
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
- except ssl.CertificateError:
+ except getattr(ssl, "CertificateError", ()):
print("REJECT")
except urllib2.URLError as exc:
if not isinstance(exc.reason, ssl.SSLError):
raise
print("REJECT")
else:
print("ACCEPT")
| Make python2-urllib2 compatible with more Python 2.7 versions | ## Code Before:
import sys
import ssl
import urllib2
if len(sys.argv) < 3 or len(sys.argv) > 4:
exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0])
host = sys.argv[1]
port = sys.argv[2]
cafile = sys.argv[3] if len(sys.argv) > 3 else None
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
except ssl.CertificateError:
print("REJECT")
except urllib2.URLError as exc:
if not isinstance(exc.reason, ssl.SSLError):
raise
print("REJECT")
else:
print("ACCEPT")
## Instruction:
Make python2-urllib2 compatible with more Python 2.7 versions
## Code After:
import sys
import ssl
import urllib2
if len(sys.argv) < 3 or len(sys.argv) > 4:
exit("Usage: %s <HOST> <PORT> [CA_FILE]" % sys.argv[0])
host = sys.argv[1]
port = sys.argv[2]
cafile = sys.argv[3] if len(sys.argv) > 3 else None
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
except getattr(ssl, "CertificateError", ()):
print("REJECT")
except urllib2.URLError as exc:
if not isinstance(exc.reason, ssl.SSLError):
raise
print("REJECT")
else:
print("ACCEPT")
| # ... existing code ...
try:
urllib2.urlopen("https://" + host + ":" + port, cafile=cafile)
except getattr(ssl, "CertificateError", ()):
print("REJECT")
except urllib2.URLError as exc:
# ... rest of the code ... |
2f67880e777c9efa5192f5c34ce5fc7d71fc0f08 | partner_communication_switzerland/wizards/end_contract_wizard.py | partner_communication_switzerland/wizards/end_contract_wizard.py |
from odoo import models, fields, api
class EndContractWizard(models.TransientModel):
_inherit = 'end.contract.wizard'
generate_communication = fields.Boolean(
'Create depart communication')
@api.multi
def end_contract(self):
self.ensure_one()
child = self.child_id
if self.generate_communication:
exit_config = self.env.ref(
'partner_communication_switzerland.'
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
default_object_ids=child.id,
default_auto_send=False).send_communication(exit_config)
return super(EndContractWizard, self).end_contract()
|
from odoo import models, fields, api
class EndContractWizard(models.TransientModel):
_inherit = 'end.contract.wizard'
generate_communication = fields.Boolean(
'Create depart communication')
@api.multi
def end_contract(self):
self.ensure_one()
if self.generate_communication:
exit_config = self.env.ref(
'partner_communication_switzerland.'
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
default_object_ids=self.contract_id.id,
default_auto_send=False).send_communication(exit_config)
return super(EndContractWizard, self).end_contract()
| FIX end contract depart letter generation | FIX end contract depart letter generation
| Python | agpl-3.0 | eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland,ecino/compassion-switzerland,ecino/compassion-switzerland,ecino/compassion-switzerland,eicher31/compassion-switzerland,eicher31/compassion-switzerland |
from odoo import models, fields, api
class EndContractWizard(models.TransientModel):
_inherit = 'end.contract.wizard'
generate_communication = fields.Boolean(
'Create depart communication')
@api.multi
def end_contract(self):
self.ensure_one()
- child = self.child_id
-
if self.generate_communication:
exit_config = self.env.ref(
'partner_communication_switzerland.'
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
- default_object_ids=child.id,
+ default_object_ids=self.contract_id.id,
default_auto_send=False).send_communication(exit_config)
return super(EndContractWizard, self).end_contract()
| FIX end contract depart letter generation | ## Code Before:
from odoo import models, fields, api
class EndContractWizard(models.TransientModel):
_inherit = 'end.contract.wizard'
generate_communication = fields.Boolean(
'Create depart communication')
@api.multi
def end_contract(self):
self.ensure_one()
child = self.child_id
if self.generate_communication:
exit_config = self.env.ref(
'partner_communication_switzerland.'
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
default_object_ids=child.id,
default_auto_send=False).send_communication(exit_config)
return super(EndContractWizard, self).end_contract()
## Instruction:
FIX end contract depart letter generation
## Code After:
from odoo import models, fields, api
class EndContractWizard(models.TransientModel):
_inherit = 'end.contract.wizard'
generate_communication = fields.Boolean(
'Create depart communication')
@api.multi
def end_contract(self):
self.ensure_one()
if self.generate_communication:
exit_config = self.env.ref(
'partner_communication_switzerland.'
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
default_object_ids=self.contract_id.id,
default_auto_send=False).send_communication(exit_config)
return super(EndContractWizard, self).end_contract()
| // ... existing code ...
def end_contract(self):
self.ensure_one()
if self.generate_communication:
exit_config = self.env.ref(
// ... modified code ...
'lifecycle_child_unplanned_exit')
self.contract_id.with_context(
default_object_ids=self.contract_id.id,
default_auto_send=False).send_communication(exit_config)
// ... rest of the code ... |
477364a4d2895fc79af2a57ace35ededf0281911 | mistral/db/sqlalchemy/migration/alembic_migrations/versions/003_cron_trigger_constraints.py | mistral/db/sqlalchemy/migration/alembic_migrations/versions/003_cron_trigger_constraints.py |
# revision identifiers, used by Alembic.
revision = '003'
down_revision = '002'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'cron_triggers_v2',
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
op.drop_index('workflow_input_hash', table_name='cron_triggers_v2')
op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2')
op.create_unique_constraint(
None,
'cron_triggers_v2', [
'workflow_input_hash', 'workflow_name', 'pattern',
'project_id', 'workflow_params_hash', 'remaining_executions',
'first_execution_time'
]
)
|
# revision identifiers, used by Alembic.
revision = '003'
down_revision = '002'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'cron_triggers_v2',
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
op.create_unique_constraint(
None,
'cron_triggers_v2', [
'workflow_input_hash', 'workflow_name', 'pattern',
'project_id', 'workflow_params_hash', 'remaining_executions',
'first_execution_time'
]
)
| Fix database upgrade from a new database | Fix database upgrade from a new database
This fixes the problem where running "mistral-db-manage upgrade heads" on a
new database result in error with workflow_input_hash index does not exist.
Change-Id: I560b2b78d11cd3fd4ae9c8606e4336e87b22ef27
Closes-Bug: #1519929
| Python | apache-2.0 | openstack/mistral,StackStorm/mistral,openstack/mistral,StackStorm/mistral |
# revision identifiers, used by Alembic.
revision = '003'
down_revision = '002'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'cron_triggers_v2',
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
+
- op.drop_index('workflow_input_hash', table_name='cron_triggers_v2')
- op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2')
op.create_unique_constraint(
None,
'cron_triggers_v2', [
'workflow_input_hash', 'workflow_name', 'pattern',
'project_id', 'workflow_params_hash', 'remaining_executions',
'first_execution_time'
]
)
| Fix database upgrade from a new database | ## Code Before:
# revision identifiers, used by Alembic.
revision = '003'
down_revision = '002'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'cron_triggers_v2',
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
op.drop_index('workflow_input_hash', table_name='cron_triggers_v2')
op.drop_index('workflow_input_hash_2', table_name='cron_triggers_v2')
op.create_unique_constraint(
None,
'cron_triggers_v2', [
'workflow_input_hash', 'workflow_name', 'pattern',
'project_id', 'workflow_params_hash', 'remaining_executions',
'first_execution_time'
]
)
## Instruction:
Fix database upgrade from a new database
## Code After:
# revision identifiers, used by Alembic.
revision = '003'
down_revision = '002'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'cron_triggers_v2',
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
op.create_unique_constraint(
None,
'cron_triggers_v2', [
'workflow_input_hash', 'workflow_name', 'pattern',
'project_id', 'workflow_params_hash', 'remaining_executions',
'first_execution_time'
]
)
| ...
sa.Column('first_execution_time', sa.DateTime(), nullable=True)
)
op.create_unique_constraint(
None,
... |
dd89173cc177f7130eca426eb4fa5737ec59c91d | test/vpp_mac.py | test/vpp_mac.py |
from util import mactobinary
class VppMacAddress():
def __init__(self, addr):
self.address = addr
def encode(self):
return {
'bytes': self.bytes
}
@property
def bytes(self):
return mactobinary(self.address)
@property
def address(self):
return self.addr.address
def __str__(self):
return self.address
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.address == other.addres
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
return self.bytes == other.bytes
else:
raise Exception("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
return False
|
from util import mactobinary
class VppMacAddress():
def __init__(self, addr):
self.address = addr
def encode(self):
return {
'bytes': self.bytes
}
@property
def bytes(self):
return mactobinary(self.address)
@property
def address(self):
return self.address
@address.setter
def address(self, value):
self.address = value
def __str__(self):
return self.address
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.address == other.address
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
return self.bytes == other.bytes
else:
raise TypeError("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
return False
| Fix L2BD arp termination Test Case | Fix L2BD arp termination Test Case
==============================================================================
L2BD arp termination Test Case
==============================================================================
12:02:21,850 Couldn't stat : /tmp/vpp-unittest-TestL2bdArpTerm-_h44qo/stats.sock
L2BD arp term - add 5 hosts, verify arp responses OK
L2BD arp term - delete 3 hosts, verify arp responses OK
L2BD arp term - recreate BD1, readd 3 hosts, verify arp responses OK
L2BD arp term - 2 IP4 addrs per host OK
L2BD arp term - create and update 10 IP4-mac pairs OK
L2BD arp/ND term - hosts with both ip4/ip6 OK
L2BD ND term - Add and Del hosts, verify ND replies OK
L2BD ND term - Add and update IP+mac, verify ND replies OK
L2BD arp term - send garps, verify arp event reports OK
L2BD arp term - send duplicate garps, verify suppression OK
L2BD arp term - disable ip4 arp events,send garps, verify no events OK
L2BD ND term - send NS packets verify reports OK
L2BD ND term - send duplicate ns, verify suppression OK
L2BD ND term - disable ip4 arp events,send ns, verify no events OK
==============================================================================
TEST RESULTS:
Scheduled tests: 14
Executed tests: 14
Passed tests: 14
==============================================================================
Test run was successful
Change-Id: I6bb1ced11b88080ffaa845d22b0bc471c4f91683
Signed-off-by: Paul Vinciguerra <[email protected]>
| Python | apache-2.0 | chrisy/vpp,vpp-dev/vpp,FDio/vpp,FDio/vpp,FDio/vpp,FDio/vpp,chrisy/vpp,FDio/vpp,chrisy/vpp,vpp-dev/vpp,chrisy/vpp,vpp-dev/vpp,vpp-dev/vpp,vpp-dev/vpp,FDio/vpp,chrisy/vpp,chrisy/vpp,chrisy/vpp,vpp-dev/vpp,vpp-dev/vpp,FDio/vpp,chrisy/vpp,FDio/vpp |
from util import mactobinary
class VppMacAddress():
def __init__(self, addr):
self.address = addr
def encode(self):
return {
'bytes': self.bytes
}
@property
def bytes(self):
return mactobinary(self.address)
@property
def address(self):
- return self.addr.address
+ return self.address
+
+ @address.setter
+ def address(self, value):
+ self.address = value
def __str__(self):
return self.address
def __eq__(self, other):
if isinstance(other, self.__class__):
- return self.address == other.addres
+ return self.address == other.address
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
return self.bytes == other.bytes
else:
- raise Exception("Comparing VppMacAddress:%s"
+ raise TypeError("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
return False
| Fix L2BD arp termination Test Case | ## Code Before:
from util import mactobinary
class VppMacAddress():
def __init__(self, addr):
self.address = addr
def encode(self):
return {
'bytes': self.bytes
}
@property
def bytes(self):
return mactobinary(self.address)
@property
def address(self):
return self.addr.address
def __str__(self):
return self.address
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.address == other.addres
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
return self.bytes == other.bytes
else:
raise Exception("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
return False
## Instruction:
Fix L2BD arp termination Test Case
## Code After:
from util import mactobinary
class VppMacAddress():
def __init__(self, addr):
self.address = addr
def encode(self):
return {
'bytes': self.bytes
}
@property
def bytes(self):
return mactobinary(self.address)
@property
def address(self):
return self.address
@address.setter
def address(self, value):
self.address = value
def __str__(self):
return self.address
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.address == other.address
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
return self.bytes == other.bytes
else:
raise TypeError("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
return False
| // ... existing code ...
@property
def address(self):
return self.address
@address.setter
def address(self, value):
self.address = value
def __str__(self):
// ... modified code ...
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.address == other.address
elif hasattr(other, "bytes"):
# vl_api_mac_addres_t
...
return self.bytes == other.bytes
else:
raise TypeError("Comparing VppMacAddress:%s"
"with unknown type: %s" %
(self, other))
// ... rest of the code ... |
b9156a522410bf39de8653bce22bb2cb56e435a0 | parktain/main.py | parktain/main.py |
from os.path import abspath, dirname, join
from gendo import Gendo
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
bot = Gendo.config_from_yaml(config_path)
@bot.listen_for('morning')
def morning(user, message):
# make sure message is "morning" and doesn't just contain it.
if message.strip() == "morning":
return "mornin' @{user.username}"
def main():
bot.run()
if __name__ == '__main__':
main()
| from os.path import abspath, dirname, join
import re
# 3rd party library
from gendo import Gendo
class Parktain(Gendo):
"""Overridden to add simple additional functionality."""
@property
def id(self):
"""Get id of the bot."""
if not hasattr(self, '_id',):
self._id = self.client.server.login_data['self']['id']
return self._id
@property
def username(self):
"""Get username of the bot."""
if not hasattr(self, '_username',):
self._username = self.client.server.username
return self.username
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
bot = Parktain.config_from_yaml(config_path)
def is_mention(f):
"""Decorator to check if bot is mentioned."""
def wrapped(name, message):
BOT_ID_RE = re.compile('<@{}>'.format(bot.id))
mention = BOT_ID_RE.search(message) is not None
if mention:
return f(name, message)
return wrapped
@bot.listen_for('morning')
def morning(user, message):
# make sure message is "morning" and doesn't just contain it.
if message.strip() == "morning":
return "mornin' @{user.username}"
@bot.listen_for('where do you live')
@is_mention
def source_code(user, message):
repo_url = 'https://github.com/punchagan/parktain'
message = 'Well, I live in your hearts...\nYou can change me from here {}, though.'
return message.format(repo_url)
def main():
bot.run()
if __name__ == '__main__':
main()
| Add 'where do you live' answer. | Add 'where do you live' answer.
| Python | bsd-3-clause | punchagan/parktain,punchagan/parktain,punchagan/parktain | + from os.path import abspath, dirname, join
+ import re
- from os.path import abspath, dirname, join
+ # 3rd party library
from gendo import Gendo
+
+
+ class Parktain(Gendo):
+ """Overridden to add simple additional functionality."""
+
+ @property
+ def id(self):
+ """Get id of the bot."""
+
+ if not hasattr(self, '_id',):
+ self._id = self.client.server.login_data['self']['id']
+ return self._id
+
+ @property
+ def username(self):
+ """Get username of the bot."""
+
+ if not hasattr(self, '_username',):
+ self._username = self.client.server.username
+ return self.username
+
+
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
- bot = Gendo.config_from_yaml(config_path)
+ bot = Parktain.config_from_yaml(config_path)
+
+ def is_mention(f):
+ """Decorator to check if bot is mentioned."""
+
+ def wrapped(name, message):
+ BOT_ID_RE = re.compile('<@{}>'.format(bot.id))
+ mention = BOT_ID_RE.search(message) is not None
+ if mention:
+ return f(name, message)
+
+ return wrapped
@bot.listen_for('morning')
def morning(user, message):
# make sure message is "morning" and doesn't just contain it.
if message.strip() == "morning":
return "mornin' @{user.username}"
+
+ @bot.listen_for('where do you live')
+ @is_mention
+ def source_code(user, message):
+ repo_url = 'https://github.com/punchagan/parktain'
+ message = 'Well, I live in your hearts...\nYou can change me from here {}, though.'
+ return message.format(repo_url)
+
+
def main():
bot.run()
if __name__ == '__main__':
main()
| Add 'where do you live' answer. | ## Code Before:
from os.path import abspath, dirname, join
from gendo import Gendo
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
bot = Gendo.config_from_yaml(config_path)
@bot.listen_for('morning')
def morning(user, message):
# make sure message is "morning" and doesn't just contain it.
if message.strip() == "morning":
return "mornin' @{user.username}"
def main():
bot.run()
if __name__ == '__main__':
main()
## Instruction:
Add 'where do you live' answer.
## Code After:
from os.path import abspath, dirname, join
import re
# 3rd party library
from gendo import Gendo
class Parktain(Gendo):
"""Overridden to add simple additional functionality."""
@property
def id(self):
"""Get id of the bot."""
if not hasattr(self, '_id',):
self._id = self.client.server.login_data['self']['id']
return self._id
@property
def username(self):
"""Get username of the bot."""
if not hasattr(self, '_username',):
self._username = self.client.server.username
return self.username
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
bot = Parktain.config_from_yaml(config_path)
def is_mention(f):
"""Decorator to check if bot is mentioned."""
def wrapped(name, message):
BOT_ID_RE = re.compile('<@{}>'.format(bot.id))
mention = BOT_ID_RE.search(message) is not None
if mention:
return f(name, message)
return wrapped
@bot.listen_for('morning')
def morning(user, message):
# make sure message is "morning" and doesn't just contain it.
if message.strip() == "morning":
return "mornin' @{user.username}"
@bot.listen_for('where do you live')
@is_mention
def source_code(user, message):
repo_url = 'https://github.com/punchagan/parktain'
message = 'Well, I live in your hearts...\nYou can change me from here {}, though.'
return message.format(repo_url)
def main():
bot.run()
if __name__ == '__main__':
main()
| ...
from os.path import abspath, dirname, join
import re
# 3rd party library
from gendo import Gendo
class Parktain(Gendo):
"""Overridden to add simple additional functionality."""
@property
def id(self):
"""Get id of the bot."""
if not hasattr(self, '_id',):
self._id = self.client.server.login_data['self']['id']
return self._id
@property
def username(self):
"""Get username of the bot."""
if not hasattr(self, '_username',):
self._username = self.client.server.username
return self.username
HERE = dirname(abspath(__file__))
config_path = join(HERE, 'config.yaml')
bot = Parktain.config_from_yaml(config_path)
def is_mention(f):
"""Decorator to check if bot is mentioned."""
def wrapped(name, message):
BOT_ID_RE = re.compile('<@{}>'.format(bot.id))
mention = BOT_ID_RE.search(message) is not None
if mention:
return f(name, message)
return wrapped
...
return "mornin' @{user.username}"
@bot.listen_for('where do you live')
@is_mention
def source_code(user, message):
repo_url = 'https://github.com/punchagan/parktain'
message = 'Well, I live in your hearts...\nYou can change me from here {}, though.'
return message.format(repo_url)
def main():
bot.run()
... |
5d9fa1838ffe7ffedb59453a0eca520b5f8d5849 | pyscf/ci/__init__.py | pyscf/ci/__init__.py | from pyscf.ci.cisd import CISD
| from pyscf.ci import cisd
def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None):
from pyscf import scf
if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)):
raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version')
return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
| Revert accidental changes to ci | Revert accidental changes to ci
| Python | apache-2.0 | gkc1000/pyscf,gkc1000/pyscf,gkc1000/pyscf,sunqm/pyscf,sunqm/pyscf,sunqm/pyscf,gkc1000/pyscf,gkc1000/pyscf,sunqm/pyscf | - from pyscf.ci.cisd import CISD
+ from pyscf.ci import cisd
+
+ def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None):
+ from pyscf import scf
+ if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)):
+ raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version')
+ return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
| Revert accidental changes to ci | ## Code Before:
from pyscf.ci.cisd import CISD
## Instruction:
Revert accidental changes to ci
## Code After:
from pyscf.ci import cisd
def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None):
from pyscf import scf
if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)):
raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version')
return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
| // ... existing code ...
from pyscf.ci import cisd
def CISD(mf, frozen=[], mo_coeff=None, mo_occ=None):
from pyscf import scf
if isinstance(mf, (scf.uhf.UHF, scf.rohf.ROHF)):
raise NotImplementedError('RO-CISD, UCISD are not available in this pyscf version')
return cisd.CISD(mf, frozen, mo_coeff, mo_occ)
// ... rest of the code ... |
0f7853c3568791f0e93ece57d2fc750dbc93b963 | starlette/concurrency.py | starlette/concurrency.py | import asyncio
import functools
import typing
from typing import Any, AsyncGenerator, Iterator
try:
import contextvars # Python 3.7+ only.
except ImportError: # pragma: no cover
contextvars = None # type: ignore
async def run_in_threadpool(
func: typing.Callable, *args: typing.Any, **kwargs: typing.Any
) -> typing.Any:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
# Ensure we run in the same context
child = functools.partial(func, *args, **kwargs)
context = contextvars.copy_context()
func = context.run
args = (child,)
elif kwargs: # pragma: no cover
# loop.run_in_executor doesn't accept 'kwargs', so bind them in here
func = functools.partial(func, **kwargs)
return await loop.run_in_executor(None, func, *args)
class _StopIteration(Exception):
pass
def _next(iterator: Iterator) -> Any:
# We can't raise `StopIteration` from within the threadpool iterator
# and catch it outside that context, so we coerce them into a different
# exception type.
try:
return next(iterator)
except StopIteration:
raise _StopIteration
async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator:
while True:
try:
yield await run_in_threadpool(_next, iterator)
except _StopIteration:
break
| import asyncio
import functools
import typing
from typing import Any, AsyncGenerator, Iterator
try:
import contextvars # Python 3.7+ only.
except ImportError: # pragma: no cover
contextvars = None # type: ignore
T = typing.TypeVar("T")
async def run_in_threadpool(
func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any
) -> T:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
# Ensure we run in the same context
child = functools.partial(func, *args, **kwargs)
context = contextvars.copy_context()
func = context.run
args = (child,)
elif kwargs: # pragma: no cover
# loop.run_in_executor doesn't accept 'kwargs', so bind them in here
func = functools.partial(func, **kwargs)
return await loop.run_in_executor(None, func, *args)
class _StopIteration(Exception):
pass
def _next(iterator: Iterator) -> Any:
# We can't raise `StopIteration` from within the threadpool iterator
# and catch it outside that context, so we coerce them into a different
# exception type.
try:
return next(iterator)
except StopIteration:
raise _StopIteration
async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator:
while True:
try:
yield await run_in_threadpool(_next, iterator)
except _StopIteration:
break
| Add type hint for run_in_threadpool return type | Add type hint for run_in_threadpool return type
| Python | bsd-3-clause | encode/starlette,encode/starlette | import asyncio
import functools
import typing
from typing import Any, AsyncGenerator, Iterator
try:
import contextvars # Python 3.7+ only.
except ImportError: # pragma: no cover
contextvars = None # type: ignore
+ T = typing.TypeVar("T")
+
async def run_in_threadpool(
- func: typing.Callable, *args: typing.Any, **kwargs: typing.Any
+ func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any
- ) -> typing.Any:
+ ) -> T:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
# Ensure we run in the same context
child = functools.partial(func, *args, **kwargs)
context = contextvars.copy_context()
func = context.run
args = (child,)
elif kwargs: # pragma: no cover
# loop.run_in_executor doesn't accept 'kwargs', so bind them in here
func = functools.partial(func, **kwargs)
return await loop.run_in_executor(None, func, *args)
class _StopIteration(Exception):
pass
def _next(iterator: Iterator) -> Any:
# We can't raise `StopIteration` from within the threadpool iterator
# and catch it outside that context, so we coerce them into a different
# exception type.
try:
return next(iterator)
except StopIteration:
raise _StopIteration
async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator:
while True:
try:
yield await run_in_threadpool(_next, iterator)
except _StopIteration:
break
| Add type hint for run_in_threadpool return type | ## Code Before:
import asyncio
import functools
import typing
from typing import Any, AsyncGenerator, Iterator
try:
import contextvars # Python 3.7+ only.
except ImportError: # pragma: no cover
contextvars = None # type: ignore
async def run_in_threadpool(
func: typing.Callable, *args: typing.Any, **kwargs: typing.Any
) -> typing.Any:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
# Ensure we run in the same context
child = functools.partial(func, *args, **kwargs)
context = contextvars.copy_context()
func = context.run
args = (child,)
elif kwargs: # pragma: no cover
# loop.run_in_executor doesn't accept 'kwargs', so bind them in here
func = functools.partial(func, **kwargs)
return await loop.run_in_executor(None, func, *args)
class _StopIteration(Exception):
pass
def _next(iterator: Iterator) -> Any:
# We can't raise `StopIteration` from within the threadpool iterator
# and catch it outside that context, so we coerce them into a different
# exception type.
try:
return next(iterator)
except StopIteration:
raise _StopIteration
async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator:
while True:
try:
yield await run_in_threadpool(_next, iterator)
except _StopIteration:
break
## Instruction:
Add type hint for run_in_threadpool return type
## Code After:
import asyncio
import functools
import typing
from typing import Any, AsyncGenerator, Iterator
try:
import contextvars # Python 3.7+ only.
except ImportError: # pragma: no cover
contextvars = None # type: ignore
T = typing.TypeVar("T")
async def run_in_threadpool(
func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any
) -> T:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
# Ensure we run in the same context
child = functools.partial(func, *args, **kwargs)
context = contextvars.copy_context()
func = context.run
args = (child,)
elif kwargs: # pragma: no cover
# loop.run_in_executor doesn't accept 'kwargs', so bind them in here
func = functools.partial(func, **kwargs)
return await loop.run_in_executor(None, func, *args)
class _StopIteration(Exception):
pass
def _next(iterator: Iterator) -> Any:
# We can't raise `StopIteration` from within the threadpool iterator
# and catch it outside that context, so we coerce them into a different
# exception type.
try:
return next(iterator)
except StopIteration:
raise _StopIteration
async def iterate_in_threadpool(iterator: Iterator) -> AsyncGenerator:
while True:
try:
yield await run_in_threadpool(_next, iterator)
except _StopIteration:
break
| // ... existing code ...
contextvars = None # type: ignore
T = typing.TypeVar("T")
async def run_in_threadpool(
func: typing.Callable[..., T], *args: typing.Any, **kwargs: typing.Any
) -> T:
loop = asyncio.get_event_loop()
if contextvars is not None: # pragma: no cover
// ... rest of the code ... |
7d02bd555d7519d485d00e02136d26a6e4e7096e | nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py | nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py |
from sqlalchemy import Column, Integer, String, MetaData, Table
meta = MetaData()
#
# Tables to alter
#
#
instance_id = Column('instance_id', Integer())
instance_uuid = Column('instance_uuid', String(255))
def upgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
migrations.c.instance_id.drop()
def downgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.c.instance_uuid.drop()
migrations.create_column(instance_id)
|
from sqlalchemy import Column, Integer, String, MetaData, Table
meta = MetaData()
#
# Tables to alter
#
#
instance_id = Column('instance_id', Integer())
instance_uuid = Column('instance_uuid', String(255))
def upgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
if migrate_engine.name == "mysql":
migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \
"`migrations_ibfk_1`;")
migrations.c.instance_id.drop()
def downgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.c.instance_uuid.drop()
migrations.create_column(instance_id)
| Drop FK before dropping instance_id column. | Drop FK before dropping instance_id column. | Python | apache-2.0 | sacharya/nova,jianghuaw/nova,leilihh/novaha,eneabio/nova,vladikr/nova_drafts,KarimAllah/nova,sileht/deb-openstack-nova,Stavitsky/nova,DirectXMan12/nova-hacking,akash1808/nova_test_latest,raildo/nova,gspilio/nova,tangfeixiong/nova,jianghuaw/nova,Juniper/nova,JioCloud/nova,zhimin711/nova,usc-isi/nova,orbitfp7/nova,JianyuWang/nova,vmturbo/nova,sebrandon1/nova,jeffrey4l/nova,Francis-Liu/animated-broccoli,psiwczak/openstack,MountainWei/nova,tianweizhang/nova,yrobla/nova,maelnor/nova,whitepages/nova,maoy/zknova,joker946/nova,russellb/nova,iuliat/nova,qwefi/nova,rahulunair/nova,berrange/nova,sileht/deb-openstack-nova,mahak/nova,fnordahl/nova,sridevikoushik31/openstack,Metaswitch/calico-nova,gooddata/openstack-nova,sebrandon1/nova,redhat-openstack/nova,eayunstack/nova,mandeepdhami/nova,tealover/nova,eharney/nova,yrobla/nova,CEG-FYP-OpenStack/scheduler,TieWei/nova,maelnor/nova,TwinkleChawla/nova,KarimAllah/nova,cloudbau/nova,isyippee/nova,mikalstill/nova,hanlind/nova,mgagne/nova,badock/nova,qwefi/nova,paulmathews/nova,kimjaejoong/nova,spring-week-topos/nova-week,plumgrid/plumgrid-nova,alaski/nova,petrutlucian94/nova,thomasem/nova,barnsnake351/nova,cernops/nova,akash1808/nova,Triv90/Nova,yrobla/nova,watonyweng/nova,akash1808/nova_test_latest,NoBodyCam/TftpPxeBootBareMetal,Tehsmash/nova,Juniper/nova,iuliat/nova,orbitfp7/nova,alexandrucoman/vbox-nova-driver,aristanetworks/arista-ovs-nova,fnordahl/nova,cernops/nova,zaina/nova,projectcalico/calico-nova,russellb/nova,apporc/nova,j-carpentier/nova,shahar-stratoscale/nova,DirectXMan12/nova-hacking,tealover/nova,vmturbo/nova,rahulunair/nova,JianyuWang/nova,varunarya10/nova_test_latest,imsplitbit/nova,klmitch/nova,silenceli/nova,NewpTone/stacklab-nova,apporc/nova,devendermishrajio/nova_test_latest,dawnpower/nova,alvarolopez/nova,felixma/nova,saleemjaveds/https-github.com-openstack-nova,adelina-t/nova,angdraug/nova,mikalstill/nova,akash1808/nova,Yuriy-Leonov/nova,CiscoSystems/nova,klmitch/nova,watonyweng/nova,devoid/nova,bgxavier/nova,citrix-openstack-build/nova,psiwczak/openstack,nikesh-mahalka/nova,sridevikoushik31/nova,CiscoSystems/nova,joker946/nova,JioCloud/nova,salv-orlando/MyRepo,rrader/nova-docker-plugin,kimjaejoong/nova,rickerc/nova_audit,savi-dev/nova,sridevikoushik31/nova,hanlind/nova,DirectXMan12/nova-hacking,blueboxgroup/nova,JioCloud/nova_test_latest,eonpatapon/nova,luogangyi/bcec-nova,belmiromoreira/nova,fajoy/nova,rickerc/nova_audit,double12gzh/nova,sileht/deb-openstack-nova,cloudbase/nova,eayunstack/nova,NeCTAR-RC/nova,aristanetworks/arista-ovs-nova,CCI-MOC/nova,sridevikoushik31/openstack,silenceli/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,virtualopensystems/nova,Juniper/nova,devendermishrajio/nova,tudorvio/nova,edulramirez/nova,bgxavier/nova,cyx1231st/nova,shootstar/novatest,varunarya10/nova_test_latest,maheshp/novatest,cernops/nova,imsplitbit/nova,maheshp/novatest,russellb/nova,josephsuh/extra-specs,mahak/nova,mgagne/nova,plumgrid/plumgrid-nova,gspilio/nova,sridevikoushik31/nova,luogangyi/bcec-nova,NoBodyCam/TftpPxeBootBareMetal,alaski/nova,cloudbau/nova,CloudServer/nova,bigswitch/nova,houshengbo/nova_vmware_compute_driver,dawnpower/nova,rajalokan/nova,belmiromoreira/nova,virtualopensystems/nova,saleemjaveds/https-github.com-openstack-nova,bclau/nova,eonpatapon/nova,Juniper/nova,citrix-openstack-build/nova,j-carpentier/nova,sacharya/nova,zhimin711/nova,Yusuke1987/openstack_template,angdraug/nova,mmnelemane/nova,eneabio/nova,cloudbase/nova,klmitch/nova,vmturbo/nova,openstack/nova,zaina/nova,edulramirez/nova,eharney/nova,josephsuh/extra-specs,cloudbase/nova,shail2810/nova,jianghuaw/nova,Triv90/Nova,NeCTAR-RC/nova,viggates/nova,zzicewind/nova,LoHChina/nova,vmturbo/nova,spring-week-topos/nova-week,noironetworks/nova,rajalokan/nova,openstack/nova,berrange/nova,takeshineshiro/nova,eneabio/nova,cloudbase/nova-virtualbox,felixma/nova,fajoy/nova,whitepages/nova,usc-isi/extra-specs,psiwczak/openstack,ruslanloman/nova,isyippee/nova,ruslanloman/nova,petrutlucian94/nova_dev,dstroppa/openstack-smartos-nova-grizzly,shahar-stratoscale/nova,bclau/nova,josephsuh/extra-specs,SUSE-Cloud/nova,vladikr/nova_drafts,noironetworks/nova,fajoy/nova,ntt-sic/nova,maoy/zknova,Francis-Liu/animated-broccoli,BeyondTheClouds/nova,blueboxgroup/nova,LoHChina/nova,cloudbase/nova-virtualbox,Triv90/Nova,jianghuaw/nova,SUSE-Cloud/nova,leilihh/novaha,devoid/nova,salv-orlando/MyRepo,Yuriy-Leonov/nova,jeffrey4l/nova,NewpTone/stacklab-nova,tangfeixiong/nova,zzicewind/nova,houshengbo/nova_vmware_compute_driver,yosshy/nova,BeyondTheClouds/nova,sridevikoushik31/openstack,aristanetworks/arista-ovs-nova,maheshp/novatest,OpenAcademy-OpenStack/nova-scheduler,mandeepdhami/nova,phenoxim/nova,paulmathews/nova,usc-isi/nova,TwinkleChawla/nova,mikalstill/nova,Metaswitch/calico-nova,ntt-sic/nova,KarimAllah/nova,houshengbo/nova_vmware_compute_driver,projectcalico/calico-nova,CloudServer/nova,savi-dev/nova,usc-isi/extra-specs,tanglei528/nova,yatinkumbhare/openstack-nova,tianweizhang/nova,Stavitsky/nova,gooddata/openstack-nova,redhat-openstack/nova,Yusuke1987/openstack_template,sridevikoushik31/nova,yatinkumbhare/openstack-nova,tanglei528/nova,leilihh/nova,dstroppa/openstack-smartos-nova-grizzly,klmitch/nova,mahak/nova,sebrandon1/nova,bigswitch/nova,rajalokan/nova,rrader/nova-docker-plugin,leilihh/nova,raildo/nova,mmnelemane/nova,ewindisch/nova,dstroppa/openstack-smartos-nova-grizzly,gooddata/openstack-nova,Tehsmash/nova,JioCloud/nova_test_latest,petrutlucian94/nova_dev,ted-gould/nova,rahulunair/nova,dims/nova,badock/nova,yosshy/nova,MountainWei/nova,scripnichenko/nova,double12gzh/nova,gspilio/nova,OpenAcademy-OpenStack/nova-scheduler,nikesh-mahalka/nova,shootstar/novatest,savi-dev/nova,scripnichenko/nova,gooddata/openstack-nova,devendermishrajio/nova_test_latest,usc-isi/extra-specs,CCI-MOC/nova,takeshineshiro/nova,usc-isi/nova,NewpTone/stacklab-nova,dims/nova,adelina-t/nova,phenoxim/nova,shail2810/nova,cyx1231st/nova,alvarolopez/nova,alexandrucoman/vbox-nova-driver,barnsnake351/nova,openstack/nova,affo/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,affo/nova,maoy/zknova,thomasem/nova,ted-gould/nova,petrutlucian94/nova,viggates/nova,CEG-FYP-OpenStack/scheduler,TieWei/nova,salv-orlando/MyRepo,hanlind/nova,rajalokan/nova,NoBodyCam/TftpPxeBootBareMetal,ewindisch/nova,BeyondTheClouds/nova,devendermishrajio/nova,paulmathews/nova,tudorvio/nova |
from sqlalchemy import Column, Integer, String, MetaData, Table
+
meta = MetaData()
#
# Tables to alter
#
#
instance_id = Column('instance_id', Integer())
instance_uuid = Column('instance_uuid', String(255))
def upgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
+
+ if migrate_engine.name == "mysql":
+ migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \
+ "`migrations_ibfk_1`;")
+
migrations.c.instance_id.drop()
def downgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.c.instance_uuid.drop()
migrations.create_column(instance_id)
| Drop FK before dropping instance_id column. | ## Code Before:
from sqlalchemy import Column, Integer, String, MetaData, Table
meta = MetaData()
#
# Tables to alter
#
#
instance_id = Column('instance_id', Integer())
instance_uuid = Column('instance_uuid', String(255))
def upgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
migrations.c.instance_id.drop()
def downgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.c.instance_uuid.drop()
migrations.create_column(instance_id)
## Instruction:
Drop FK before dropping instance_id column.
## Code After:
from sqlalchemy import Column, Integer, String, MetaData, Table
meta = MetaData()
#
# Tables to alter
#
#
instance_id = Column('instance_id', Integer())
instance_uuid = Column('instance_uuid', String(255))
def upgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
if migrate_engine.name == "mysql":
migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \
"`migrations_ibfk_1`;")
migrations.c.instance_id.drop()
def downgrade(migrate_engine):
meta.bind = migrate_engine
migrations = Table('migrations', meta, autoload=True)
migrations.c.instance_uuid.drop()
migrations.create_column(instance_id)
| # ... existing code ...
from sqlalchemy import Column, Integer, String, MetaData, Table
meta = MetaData()
# ... modified code ...
migrations = Table('migrations', meta, autoload=True)
migrations.create_column(instance_uuid)
if migrate_engine.name == "mysql":
migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \
"`migrations_ibfk_1`;")
migrations.c.instance_id.drop()
# ... rest of the code ... |
23072e882edb6da55cb12ef0591a786235249670 | ome/__main__.py | ome/__main__.py |
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
| Use print_verbose for conditional printing. | Use print_verbose for conditional printing.
| Python | mit | shaurz/ome,shaurz/ome |
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
+
+ def print_verbose(*args, **kwargs):
+ if command_args.verbose:
+ print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
- if command_args.verbose:
- print('ome: using target {}'.format(target.name))
+ print_verbose('ome: using target {}'.format(target.name))
- print('ome: using backend {} {}'.format(backend.name, backend.version))
+ print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
- if command_args.verbose:
- print('ome: compiling {}'.format(filename))
+ print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
| Use print_verbose for conditional printing. | ## Code Before:
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
## Instruction:
Use print_verbose for conditional printing.
## Code After:
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
| // ... existing code ...
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
// ... modified code ...
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
// ... rest of the code ... |
4b43a2f50740bbeab95f64137eb8993ed8ac4617 | other/password_generator.py | other/password_generator.py | import string
from random import *
letters = string.ascii_letters
digits = string.digits
symbols = string.punctuation
chars = letters + digits + symbols
min_length = 8
max_length = 16
password = ''.join(choice(chars) for x in range(randint(min_length, max_length)))
print('Password: %s' % password)
print('[ If you are thinking of using this passsword, You better save it. ]')
| import string
import random
letters = [letter for letter in string.ascii_letters]
digits = [digit for digit in string.digits]
symbols = [symbol for symbol in string.punctuation]
chars = letters + digits + symbols
random.shuffle(chars)
min_length = 8
max_length = 16
password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length)))
print('Password: ' + password)
print('[ If you are thinking of using this passsword, You better save it. ]')
| Add another randomness into the password generator | Add another randomness into the password generator
Uses import random for namespace cleanliness
Uses list instead of string for 'chars' variable in order to shuffle, increases randomness
Instead of string formatting, uses string concatenation because (currently) it is simpler | Python | mit | TheAlgorithms/Python | import string
- from random import *
+ import random
- letters = string.ascii_letters
- digits = string.digits
- symbols = string.punctuation
+ letters = [letter for letter in string.ascii_letters]
+ digits = [digit for digit in string.digits]
+ symbols = [symbol for symbol in string.punctuation]
chars = letters + digits + symbols
+ random.shuffle(chars)
min_length = 8
max_length = 16
- password = ''.join(choice(chars) for x in range(randint(min_length, max_length)))
+ password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length)))
- print('Password: %s' % password)
+ print('Password: ' + password)
print('[ If you are thinking of using this passsword, You better save it. ]')
| Add another randomness into the password generator | ## Code Before:
import string
from random import *
letters = string.ascii_letters
digits = string.digits
symbols = string.punctuation
chars = letters + digits + symbols
min_length = 8
max_length = 16
password = ''.join(choice(chars) for x in range(randint(min_length, max_length)))
print('Password: %s' % password)
print('[ If you are thinking of using this passsword, You better save it. ]')
## Instruction:
Add another randomness into the password generator
## Code After:
import string
import random
letters = [letter for letter in string.ascii_letters]
digits = [digit for digit in string.digits]
symbols = [symbol for symbol in string.punctuation]
chars = letters + digits + symbols
random.shuffle(chars)
min_length = 8
max_length = 16
password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length)))
print('Password: ' + password)
print('[ If you are thinking of using this passsword, You better save it. ]')
| ...
import string
import random
letters = [letter for letter in string.ascii_letters]
digits = [digit for digit in string.digits]
symbols = [symbol for symbol in string.punctuation]
chars = letters + digits + symbols
random.shuffle(chars)
min_length = 8
max_length = 16
password = ''.join(random.choice(chars) for x in range(random.randint(min_length, max_length)))
print('Password: ' + password)
print('[ If you are thinking of using this passsword, You better save it. ]')
... |
ba98874be9370ec49c2c04e89d456f723b5d083c | monitoring/test/test_data/exceptions.py | monitoring/test/test_data/exceptions.py |
from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
def data(TEST):
TEST.exceptions = exceptions.data
monitoring_exception = monascacli.ClientException
TEST.exceptions.monitoring = exceptions.create_stubbed_exception(
monitoring_exception)
| try:
from monascaclient.apiclient import exceptions as monascacli
except ImportError:
from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
def data(TEST):
TEST.exceptions = exceptions.data
monitoring_exception = monascacli.ClientException
TEST.exceptions.monitoring = exceptions.create_stubbed_exception(
monitoring_exception)
| Adjust tests for python-monascaclient >= 1.3.0 | Adjust tests for python-monascaclient >= 1.3.0
the exceptions module was moved out of the openstack.common namespace,
so try to import the new location first and fall back to the old
one if it doesn't exist.
Change-Id: I3305775baaab15dca8d5e7e5cfc0932f94d4d153
| Python | apache-2.0 | openstack/monasca-ui,openstack/monasca-ui,openstack/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,stackforge/monasca-ui,openstack/monasca-ui | + try:
+ from monascaclient.apiclient import exceptions as monascacli
+ except ImportError:
+ from monascaclient.openstack.common.apiclient import exceptions as monascacli
- from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
def data(TEST):
TEST.exceptions = exceptions.data
monitoring_exception = monascacli.ClientException
TEST.exceptions.monitoring = exceptions.create_stubbed_exception(
monitoring_exception)
| Adjust tests for python-monascaclient >= 1.3.0 | ## Code Before:
from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
def data(TEST):
TEST.exceptions = exceptions.data
monitoring_exception = monascacli.ClientException
TEST.exceptions.monitoring = exceptions.create_stubbed_exception(
monitoring_exception)
## Instruction:
Adjust tests for python-monascaclient >= 1.3.0
## Code After:
try:
from monascaclient.apiclient import exceptions as monascacli
except ImportError:
from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
def data(TEST):
TEST.exceptions = exceptions.data
monitoring_exception = monascacli.ClientException
TEST.exceptions.monitoring = exceptions.create_stubbed_exception(
monitoring_exception)
| # ... existing code ...
try:
from monascaclient.apiclient import exceptions as monascacli
except ImportError:
from monascaclient.openstack.common.apiclient import exceptions as monascacli
from openstack_dashboard.test.test_data import exceptions
# ... rest of the code ... |
b2b1c2b8543cae37990262b2a811a9b0f26327da | arm/utils/locker.py | arm/utils/locker.py | from kvs import CacheKvs
class Locker(object):
"""
locker for move the locker
"""
LOCKER_KEY = 'locker'
EXPIRES = 5 # 5 sec
def __init__(self, key=None):
self.key = self.LOCKER_KEY
if key:
self.key += '.{}'.format(key)
self.locker = CacheKvs(self.key)
def lock(self):
self.locker.set('locked', expires=self.EXPIRES, nx=True)
def unlock(self):
self.locker.delete()
def is_lock(self):
return self.locker.get() == 'locked'
def on_lock(self, func):
def wrapper(*args, **kwargs):
if self.is_lock():
return
self.lock()
try:
return func(*args, **kwargs)
except Exception as e:
raise e
finally:
self.unlock()
return wrapper
| from kvs import CacheKvs
class Locker(object):
"""
locker for move the locker
"""
LOCKER_KEY = 'locker'
EXPIRES = 5 # 5 sec
def __init__(self, key=None):
self.key = self.LOCKER_KEY
if key:
self.key += '.{}'.format(key)
self.locker = CacheKvs(self.key)
def lock(self):
self.locker.set('locked', expires=self.EXPIRES, nx=True)
def unlock(self):
self.locker.delete()
def is_lock(self):
return self.locker.get() == 'locked'
def on_lock(self, func):
def wrapper(*args, **kwargs):
if self.lock():
try:
return func(*args, **kwargs)
except Exception as e:
raise e
finally:
self.unlock()
return wrapper
| Fix redis lock, use SETNX | Fix redis lock, use SETNX
| Python | mit | mapler/tuesday,mapler/tuesday,mapler/tuesday | from kvs import CacheKvs
class Locker(object):
"""
locker for move the locker
"""
LOCKER_KEY = 'locker'
EXPIRES = 5 # 5 sec
def __init__(self, key=None):
self.key = self.LOCKER_KEY
if key:
self.key += '.{}'.format(key)
self.locker = CacheKvs(self.key)
def lock(self):
self.locker.set('locked', expires=self.EXPIRES, nx=True)
def unlock(self):
self.locker.delete()
def is_lock(self):
return self.locker.get() == 'locked'
def on_lock(self, func):
def wrapper(*args, **kwargs):
- if self.is_lock():
+ if self.lock():
- return
- self.lock()
- try:
+ try:
- return func(*args, **kwargs)
+ return func(*args, **kwargs)
- except Exception as e:
+ except Exception as e:
- raise e
+ raise e
- finally:
+ finally:
- self.unlock()
+ self.unlock()
return wrapper
| Fix redis lock, use SETNX | ## Code Before:
from kvs import CacheKvs
class Locker(object):
"""
locker for move the locker
"""
LOCKER_KEY = 'locker'
EXPIRES = 5 # 5 sec
def __init__(self, key=None):
self.key = self.LOCKER_KEY
if key:
self.key += '.{}'.format(key)
self.locker = CacheKvs(self.key)
def lock(self):
self.locker.set('locked', expires=self.EXPIRES, nx=True)
def unlock(self):
self.locker.delete()
def is_lock(self):
return self.locker.get() == 'locked'
def on_lock(self, func):
def wrapper(*args, **kwargs):
if self.is_lock():
return
self.lock()
try:
return func(*args, **kwargs)
except Exception as e:
raise e
finally:
self.unlock()
return wrapper
## Instruction:
Fix redis lock, use SETNX
## Code After:
from kvs import CacheKvs
class Locker(object):
"""
locker for move the locker
"""
LOCKER_KEY = 'locker'
EXPIRES = 5 # 5 sec
def __init__(self, key=None):
self.key = self.LOCKER_KEY
if key:
self.key += '.{}'.format(key)
self.locker = CacheKvs(self.key)
def lock(self):
self.locker.set('locked', expires=self.EXPIRES, nx=True)
def unlock(self):
self.locker.delete()
def is_lock(self):
return self.locker.get() == 'locked'
def on_lock(self, func):
def wrapper(*args, **kwargs):
if self.lock():
try:
return func(*args, **kwargs)
except Exception as e:
raise e
finally:
self.unlock()
return wrapper
| // ... existing code ...
def on_lock(self, func):
def wrapper(*args, **kwargs):
if self.lock():
try:
return func(*args, **kwargs)
except Exception as e:
raise e
finally:
self.unlock()
return wrapper
// ... rest of the code ... |
d1628356c7981748e2446c7b43d33d21cdef7e02 | geoengine_partner/geo_partner.py | geoengine_partner/geo_partner.py |
from openerp.osv import fields
from base_geoengine import geo_model
class ResPartner(geo_model.GeoModel):
"""Add geo_point to partner using a function filed"""
_name = "res.partner"
_inherit = "res.partner"
_columns = {
'geo_point': fields.geo_point('Addresses coordinate')
}
|
from openerp.osv import fields
from openerp.addons.base_geoengine import geo_model
class ResPartner(geo_model.GeoModel):
"""Add geo_point to partner using a function filed"""
_name = "res.partner"
_inherit = "res.partner"
_columns = {
'geo_point': fields.geo_point('Addresses coordinate')
}
| Use absolute imports on opnerp.addons | [FIX] Use absolute imports on opnerp.addons
| Python | agpl-3.0 | OCA/geospatial,OCA/geospatial,OCA/geospatial |
from openerp.osv import fields
- from base_geoengine import geo_model
+ from openerp.addons.base_geoengine import geo_model
class ResPartner(geo_model.GeoModel):
"""Add geo_point to partner using a function filed"""
_name = "res.partner"
_inherit = "res.partner"
_columns = {
'geo_point': fields.geo_point('Addresses coordinate')
}
| Use absolute imports on opnerp.addons | ## Code Before:
from openerp.osv import fields
from base_geoengine import geo_model
class ResPartner(geo_model.GeoModel):
"""Add geo_point to partner using a function filed"""
_name = "res.partner"
_inherit = "res.partner"
_columns = {
'geo_point': fields.geo_point('Addresses coordinate')
}
## Instruction:
Use absolute imports on opnerp.addons
## Code After:
from openerp.osv import fields
from openerp.addons.base_geoengine import geo_model
class ResPartner(geo_model.GeoModel):
"""Add geo_point to partner using a function filed"""
_name = "res.partner"
_inherit = "res.partner"
_columns = {
'geo_point': fields.geo_point('Addresses coordinate')
}
| ...
from openerp.osv import fields
from openerp.addons.base_geoengine import geo_model
... |
8b07dde78e753f6dce663481a68856024ed2fc49 | plutokore/__init__.py | plutokore/__init__.py | from .environments.makino import MakinoProfile
from .environments.king import KingProfile
from .jet import AstroJet
from . import luminosity
from . import plotting
from . import simulations
from . import helpers
from . import io
__all__ = [
'environments',
'luminosity',
'plotting',
'simulations',
'jet',
'helpers',
'io',
]
| from .environments.makino import MakinoProfile
from .environments.king import KingProfile
from .jet import AstroJet
from . import luminosity
from . import plotting
from . import simulations
from . import helpers
from . import io
from . import configuration
__all__ = [
'environments',
'luminosity',
'plotting',
'simulations',
'jet',
'helpers',
'io',
'configuration',
]
| Add configuration module to package exports | Add configuration module to package exports
| Python | mit | opcon/plutokore,opcon/plutokore | from .environments.makino import MakinoProfile
from .environments.king import KingProfile
from .jet import AstroJet
from . import luminosity
from . import plotting
from . import simulations
from . import helpers
from . import io
+ from . import configuration
__all__ = [
'environments',
'luminosity',
'plotting',
'simulations',
'jet',
'helpers',
'io',
+ 'configuration',
]
| Add configuration module to package exports | ## Code Before:
from .environments.makino import MakinoProfile
from .environments.king import KingProfile
from .jet import AstroJet
from . import luminosity
from . import plotting
from . import simulations
from . import helpers
from . import io
__all__ = [
'environments',
'luminosity',
'plotting',
'simulations',
'jet',
'helpers',
'io',
]
## Instruction:
Add configuration module to package exports
## Code After:
from .environments.makino import MakinoProfile
from .environments.king import KingProfile
from .jet import AstroJet
from . import luminosity
from . import plotting
from . import simulations
from . import helpers
from . import io
from . import configuration
__all__ = [
'environments',
'luminosity',
'plotting',
'simulations',
'jet',
'helpers',
'io',
'configuration',
]
| # ... existing code ...
from . import helpers
from . import io
from . import configuration
__all__ = [
# ... modified code ...
'helpers',
'io',
'configuration',
]
# ... rest of the code ... |
01036133ed749d96a74bafb6b3f8670c06c63a84 | 1selfOpenDashboardCommand.py | 1selfOpenDashboardCommand.py | import sublime, sublime_plugin, webbrowser
QD_URL = "https://app.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
def run(self,edit):
SETTINGS = {}
SETTINGS_FILE = "1self.sublime-settings"
SETTINGS = sublime.load_settings(SETTINGS_FILE)
stream_id = SETTINGS.get("streamId")
read_token = SETTINGS.get("readToken")
VERSION = SETTINGS.get("VERSION")
qd_url = QD_URL
url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url) | import sublime, sublime_plugin, webbrowser
QD_URL = "http://www.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
def run(self,edit):
SETTINGS = {}
SETTINGS_FILE = "1self.sublime-settings"
SETTINGS = sublime.load_settings(SETTINGS_FILE)
stream_id = SETTINGS.get("streamId")
read_token = SETTINGS.get("readToken")
VERSION = SETTINGS.get("VERSION")
qd_url = QD_URL
url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url) | Change landing URLs to website | Change landing URLs to website
| Python | apache-2.0 | 1self/sublime-text-plugin,1self/sublime-text-plugin,1self/sublime-text-plugin | import sublime, sublime_plugin, webbrowser
- QD_URL = "https://app.1self.co"
+ QD_URL = "http://www.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
def run(self,edit):
SETTINGS = {}
SETTINGS_FILE = "1self.sublime-settings"
SETTINGS = sublime.load_settings(SETTINGS_FILE)
stream_id = SETTINGS.get("streamId")
read_token = SETTINGS.get("readToken")
VERSION = SETTINGS.get("VERSION")
qd_url = QD_URL
- url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
+ url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url) | Change landing URLs to website | ## Code Before:
import sublime, sublime_plugin, webbrowser
QD_URL = "https://app.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
def run(self,edit):
SETTINGS = {}
SETTINGS_FILE = "1self.sublime-settings"
SETTINGS = sublime.load_settings(SETTINGS_FILE)
stream_id = SETTINGS.get("streamId")
read_token = SETTINGS.get("readToken")
VERSION = SETTINGS.get("VERSION")
qd_url = QD_URL
url = "%(qd_url)s/dashboard?streamId=%(stream_id)s&readToken=%(read_token)s&source=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url)
## Instruction:
Change landing URLs to website
## Code After:
import sublime, sublime_plugin, webbrowser
QD_URL = "http://www.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
def run(self,edit):
SETTINGS = {}
SETTINGS_FILE = "1self.sublime-settings"
SETTINGS = sublime.load_settings(SETTINGS_FILE)
stream_id = SETTINGS.get("streamId")
read_token = SETTINGS.get("readToken")
VERSION = SETTINGS.get("VERSION")
qd_url = QD_URL
url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url) | ...
QD_URL = "http://www.1self.co"
class GoTo1selfDashboardCommand(sublime_plugin.TextCommand):
...
qd_url = QD_URL
url = "%(qd_url)s/?streamid=%(stream_id)s&readToken=%(read_token)s&appid=app-id-598358b6aacda229634d443c9539662b&version=%(VERSION)s" % locals()
print(url)
webbrowser.open_new_tab(url)
... |
668a5240c29047d86fe9451f3078bb163bea0db9 | skan/__init__.py | skan/__init__.py | from .csr import skeleton_to_csgraph, branch_statistics, summarise
__all__ = ['skeleton_to_csgraph',
'branch_statistics',
'summarise'] | from .csr import skeleton_to_csgraph, branch_statistics, summarise
__version__ = '0.1-dev'
__all__ = ['skeleton_to_csgraph',
'branch_statistics',
'summarise']
| Add version info to package init | Add version info to package init
| Python | bsd-3-clause | jni/skan | from .csr import skeleton_to_csgraph, branch_statistics, summarise
+
+ __version__ = '0.1-dev'
__all__ = ['skeleton_to_csgraph',
'branch_statistics',
'summarise']
+ | Add version info to package init | ## Code Before:
from .csr import skeleton_to_csgraph, branch_statistics, summarise
__all__ = ['skeleton_to_csgraph',
'branch_statistics',
'summarise']
## Instruction:
Add version info to package init
## Code After:
from .csr import skeleton_to_csgraph, branch_statistics, summarise
__version__ = '0.1-dev'
__all__ = ['skeleton_to_csgraph',
'branch_statistics',
'summarise']
| # ... existing code ...
from .csr import skeleton_to_csgraph, branch_statistics, summarise
__version__ = '0.1-dev'
__all__ = ['skeleton_to_csgraph',
# ... rest of the code ... |
b5f980b700707ecc611746f93b1f62650c76c451 | pgcrypto_fields/aggregates.py | pgcrypto_fields/aggregates.py | from django.db import models
class Decrypt(models.Aggregate):
"""`Decrypt` creates an alias for `DecryptFunctionSQL`.
`alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`.
`self.lookup` is defined in `models.Aggregate.__init__`.
"""
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
from pgcrypto_fields.sql import aggregates
klass = getattr(aggregates, self.name)
aggregate = klass(
col,
source=source,
is_summary=is_summary,
**self.extra
)
query.aggregates[alias] = aggregate
class PGPPub(Decrypt):
"""PGP public key based aggregation."""
name = 'PGPPub'
class PGPSym(Decrypt):
"""PGP symmetric key based aggregation."""
name = 'PGPSym'
| from django.db import models
from pgcrypto_fields.sql import aggregates
class Decrypt(models.Aggregate):
"""`Decrypt` creates an alias for `DecryptFunctionSQL`.
`alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`.
`self.lookup` is defined in `models.Aggregate.__init__`.
"""
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
klass = getattr(aggregates, self.name)
aggregate = klass(
col,
source=source,
is_summary=is_summary,
**self.extra
)
query.aggregates[alias] = aggregate
class PGPPub(Decrypt):
"""PGP public key based aggregation."""
name = 'PGPPub'
class PGPSym(Decrypt):
"""PGP symmetric key based aggregation."""
name = 'PGPSym'
| Move import to top of the file | Move import to top of the file
| Python | bsd-2-clause | incuna/django-pgcrypto-fields,atdsaa/django-pgcrypto-fields | from django.db import models
+
+ from pgcrypto_fields.sql import aggregates
class Decrypt(models.Aggregate):
"""`Decrypt` creates an alias for `DecryptFunctionSQL`.
`alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`.
`self.lookup` is defined in `models.Aggregate.__init__`.
"""
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
- from pgcrypto_fields.sql import aggregates
-
klass = getattr(aggregates, self.name)
aggregate = klass(
col,
source=source,
is_summary=is_summary,
**self.extra
)
query.aggregates[alias] = aggregate
class PGPPub(Decrypt):
"""PGP public key based aggregation."""
name = 'PGPPub'
class PGPSym(Decrypt):
"""PGP symmetric key based aggregation."""
name = 'PGPSym'
| Move import to top of the file | ## Code Before:
from django.db import models
class Decrypt(models.Aggregate):
"""`Decrypt` creates an alias for `DecryptFunctionSQL`.
`alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`.
`self.lookup` is defined in `models.Aggregate.__init__`.
"""
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
from pgcrypto_fields.sql import aggregates
klass = getattr(aggregates, self.name)
aggregate = klass(
col,
source=source,
is_summary=is_summary,
**self.extra
)
query.aggregates[alias] = aggregate
class PGPPub(Decrypt):
"""PGP public key based aggregation."""
name = 'PGPPub'
class PGPSym(Decrypt):
"""PGP symmetric key based aggregation."""
name = 'PGPSym'
## Instruction:
Move import to top of the file
## Code After:
from django.db import models
from pgcrypto_fields.sql import aggregates
class Decrypt(models.Aggregate):
"""`Decrypt` creates an alias for `DecryptFunctionSQL`.
`alias` is `{self.lookup}__decrypt` where 'decrypt' is `self.name.lower()`.
`self.lookup` is defined in `models.Aggregate.__init__`.
"""
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
klass = getattr(aggregates, self.name)
aggregate = klass(
col,
source=source,
is_summary=is_summary,
**self.extra
)
query.aggregates[alias] = aggregate
class PGPPub(Decrypt):
"""PGP public key based aggregation."""
name = 'PGPPub'
class PGPSym(Decrypt):
"""PGP symmetric key based aggregation."""
name = 'PGPSym'
| # ... existing code ...
from django.db import models
from pgcrypto_fields.sql import aggregates
# ... modified code ...
def add_to_query(self, query, alias, col, source, is_summary):
"""Add the aggregate to the query."""
klass = getattr(aggregates, self.name)
aggregate = klass(
# ... rest of the code ... |
7fc4e7382665cf9eac4d19efcf9641ad57271e87 | organizer/models.py | organizer/models.py | from django.db import models
# Model Field Reference
# https://docs.djangoproject.com/en/1.8/ref/models/fields/
class Tag(models.Model):
name = models.CharField(
max_length=31, unique=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
class Meta:
ordering = ['name']
def __str__(self):
return self.name
class Startup(models.Model):
name = models.CharField(
max_length=31, db_index=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
description = models.TextField()
founded_date = models.DateField(
'date founded')
contact = models.EmailField()
website = models.URLField(max_length=255)
tags = models.ManyToManyField(Tag)
def __str__(self):
return self.name
class NewsLink(models.Model):
title = models.CharField(max_length=63)
pub_date = models.DateField('date published')
link = models.URLField(max_length=255)
startup = models.ForeignKey(Startup)
def __str__(self):
return "{}: {}".format(
self.startup, self.title)
| from django.db import models
# Model Field Reference
# https://docs.djangoproject.com/en/1.8/ref/models/fields/
class Tag(models.Model):
name = models.CharField(
max_length=31, unique=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
class Meta:
ordering = ['name']
def __str__(self):
return self.name
class Startup(models.Model):
name = models.CharField(
max_length=31, db_index=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
description = models.TextField()
founded_date = models.DateField(
'date founded')
contact = models.EmailField()
website = models.URLField(max_length=255)
tags = models.ManyToManyField(Tag)
class Meta:
ordering = ['name']
get_latest_by = 'founded_date'
def __str__(self):
return self.name
class NewsLink(models.Model):
title = models.CharField(max_length=63)
pub_date = models.DateField('date published')
link = models.URLField(max_length=255)
startup = models.ForeignKey(Startup)
def __str__(self):
return "{}: {}".format(
self.startup, self.title)
| Declare Meta class in Startup model. | Ch03: Declare Meta class in Startup model. [skip ci]
| Python | bsd-2-clause | jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8 | from django.db import models
# Model Field Reference
# https://docs.djangoproject.com/en/1.8/ref/models/fields/
class Tag(models.Model):
name = models.CharField(
max_length=31, unique=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
class Meta:
ordering = ['name']
def __str__(self):
return self.name
class Startup(models.Model):
name = models.CharField(
max_length=31, db_index=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
description = models.TextField()
founded_date = models.DateField(
'date founded')
contact = models.EmailField()
website = models.URLField(max_length=255)
tags = models.ManyToManyField(Tag)
+ class Meta:
+ ordering = ['name']
+ get_latest_by = 'founded_date'
+
def __str__(self):
return self.name
class NewsLink(models.Model):
title = models.CharField(max_length=63)
pub_date = models.DateField('date published')
link = models.URLField(max_length=255)
startup = models.ForeignKey(Startup)
def __str__(self):
return "{}: {}".format(
self.startup, self.title)
| Declare Meta class in Startup model. | ## Code Before:
from django.db import models
# Model Field Reference
# https://docs.djangoproject.com/en/1.8/ref/models/fields/
class Tag(models.Model):
name = models.CharField(
max_length=31, unique=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
class Meta:
ordering = ['name']
def __str__(self):
return self.name
class Startup(models.Model):
name = models.CharField(
max_length=31, db_index=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
description = models.TextField()
founded_date = models.DateField(
'date founded')
contact = models.EmailField()
website = models.URLField(max_length=255)
tags = models.ManyToManyField(Tag)
def __str__(self):
return self.name
class NewsLink(models.Model):
title = models.CharField(max_length=63)
pub_date = models.DateField('date published')
link = models.URLField(max_length=255)
startup = models.ForeignKey(Startup)
def __str__(self):
return "{}: {}".format(
self.startup, self.title)
## Instruction:
Declare Meta class in Startup model.
## Code After:
from django.db import models
# Model Field Reference
# https://docs.djangoproject.com/en/1.8/ref/models/fields/
class Tag(models.Model):
name = models.CharField(
max_length=31, unique=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
class Meta:
ordering = ['name']
def __str__(self):
return self.name
class Startup(models.Model):
name = models.CharField(
max_length=31, db_index=True)
slug = models.SlugField(
max_length=31,
unique=True,
help_text='A label for URL config.')
description = models.TextField()
founded_date = models.DateField(
'date founded')
contact = models.EmailField()
website = models.URLField(max_length=255)
tags = models.ManyToManyField(Tag)
class Meta:
ordering = ['name']
get_latest_by = 'founded_date'
def __str__(self):
return self.name
class NewsLink(models.Model):
title = models.CharField(max_length=63)
pub_date = models.DateField('date published')
link = models.URLField(max_length=255)
startup = models.ForeignKey(Startup)
def __str__(self):
return "{}: {}".format(
self.startup, self.title)
| # ... existing code ...
tags = models.ManyToManyField(Tag)
class Meta:
ordering = ['name']
get_latest_by = 'founded_date'
def __str__(self):
return self.name
# ... rest of the code ... |
d44010acc32fcb78570cd34478d0f4e8f1cfa979 | utility/dbproc.py | utility/dbproc.py | from discord.ext import commands
from utils import *
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
from member import Base, Member
import discord
import asyncio
class Baydb:
engine = create_engine('sqlite:///bayohwoolph.db')
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
session = DBSession()
conn = engine.connect()
| from discord.ext import commands
from utils import *
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
from member import Base, Member
from config import Config
import discord
import asyncio
class Baydb:
engine = create_engine(Config.MAIN['dbpath'])
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
session = DBSession()
conn = engine.connect()
| Move another usage of DB into ini file thing. | Move another usage of DB into ini file thing.
| Python | agpl-3.0 | dark-echo/Bay-Oh-Woolph,freiheit/Bay-Oh-Woolph | from discord.ext import commands
from utils import *
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
from member import Base, Member
+ from config import Config
import discord
import asyncio
class Baydb:
- engine = create_engine('sqlite:///bayohwoolph.db')
+ engine = create_engine(Config.MAIN['dbpath'])
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
session = DBSession()
conn = engine.connect()
| Move another usage of DB into ini file thing. | ## Code Before:
from discord.ext import commands
from utils import *
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
from member import Base, Member
import discord
import asyncio
class Baydb:
engine = create_engine('sqlite:///bayohwoolph.db')
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
session = DBSession()
conn = engine.connect()
## Instruction:
Move another usage of DB into ini file thing.
## Code After:
from discord.ext import commands
from utils import *
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
from member import Base, Member
from config import Config
import discord
import asyncio
class Baydb:
engine = create_engine(Config.MAIN['dbpath'])
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
session = DBSession()
conn = engine.connect()
| // ... existing code ...
from sqlalchemy.orm import sessionmaker
from member import Base, Member
from config import Config
import discord
import asyncio
// ... modified code ...
class Baydb:
engine = create_engine(Config.MAIN['dbpath'])
Base.metadata.bind = engine
DBSession = sessionmaker(bind=engine)
// ... rest of the code ... |
82dcd51c59eecccac4e7d9ee1dac754b27ff9ed2 | mzalendo/feedback/views.py | mzalendo/feedback/views.py | from django.shortcuts import render_to_response
from django.template import RequestContext
from django.views.decorators.csrf import csrf_protect
from models import Feedback
from forms import FeedbackForm
@csrf_protect
def add(request):
"""Gather feedback for a page, and if it is ok show a thanks message and link back to the page."""
submit_was_success = False
return_to_url = None
# If it is a post request try to create the feedback
if request.method == 'POST':
form = FeedbackForm( request.POST )
if form.is_valid():
feedback = Feedback()
feedback.url = form.cleaned_data['url']
feedback.email = form.cleaned_data['email']
feedback.comment = form.cleaned_data['comment']
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
feedback.status = 'spammy'
if request.user.is_authenticated():
feedback.user = request.user
feedback.save()
submit_was_success = True
return_to_url = feedback.url or None
else:
# use GET to grab the url if set
form = FeedbackForm(initial=request.GET)
return render_to_response(
'feedback/add.html',
{
'form': form,
'submit_was_success': submit_was_success,
'return_to_url': return_to_url,
},
context_instance=RequestContext(request)
)
| from django.shortcuts import render_to_response
from django.template import RequestContext
from django.views.decorators.csrf import csrf_protect
from models import Feedback
from forms import FeedbackForm
import re
@csrf_protect
def add(request):
"""Gather feedback for a page, and if it is ok show a thanks message and link back to the page."""
submit_was_success = False
return_to_url = None
# If it is a post request try to create the feedback
if request.method == 'POST':
form = FeedbackForm( request.POST )
if form.is_valid():
feedback = Feedback()
feedback.url = form.cleaned_data['url']
feedback.email = form.cleaned_data['email']
feedback.comment = form.cleaned_data['comment']
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
feedback.status = 'spammy'
# if the comment starts with an html tag it is probably spam
if re.search('\A\s*<\w+>', form.cleaned_data['comment']):
feedback.status = 'spammy'
if request.user.is_authenticated():
feedback.user = request.user
feedback.save()
submit_was_success = True
return_to_url = feedback.url or None
else:
# use GET to grab the url if set
form = FeedbackForm(initial=request.GET)
return render_to_response(
'feedback/add.html',
{
'form': form,
'submit_was_success': submit_was_success,
'return_to_url': return_to_url,
},
context_instance=RequestContext(request)
)
| Mark feedback comments starting with a html tag as spammy | Mark feedback comments starting with a html tag as spammy
| Python | agpl-3.0 | ken-muturi/pombola,mysociety/pombola,hzj123/56th,hzj123/56th,mysociety/pombola,ken-muturi/pombola,hzj123/56th,Hutspace/odekro,geoffkilpin/pombola,patricmutwiri/pombola,mysociety/pombola,patricmutwiri/pombola,patricmutwiri/pombola,hzj123/56th,Hutspace/odekro,ken-muturi/pombola,Hutspace/odekro,patricmutwiri/pombola,Hutspace/odekro,mysociety/pombola,ken-muturi/pombola,geoffkilpin/pombola,geoffkilpin/pombola,patricmutwiri/pombola,hzj123/56th,geoffkilpin/pombola,ken-muturi/pombola,mysociety/pombola,patricmutwiri/pombola,geoffkilpin/pombola,hzj123/56th,Hutspace/odekro,geoffkilpin/pombola,mysociety/pombola,ken-muturi/pombola | from django.shortcuts import render_to_response
from django.template import RequestContext
from django.views.decorators.csrf import csrf_protect
from models import Feedback
from forms import FeedbackForm
+
+ import re
+
@csrf_protect
def add(request):
"""Gather feedback for a page, and if it is ok show a thanks message and link back to the page."""
submit_was_success = False
return_to_url = None
# If it is a post request try to create the feedback
if request.method == 'POST':
form = FeedbackForm( request.POST )
if form.is_valid():
feedback = Feedback()
feedback.url = form.cleaned_data['url']
feedback.email = form.cleaned_data['email']
feedback.comment = form.cleaned_data['comment']
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
+ feedback.status = 'spammy'
+
+ # if the comment starts with an html tag it is probably spam
+ if re.search('\A\s*<\w+>', form.cleaned_data['comment']):
feedback.status = 'spammy'
if request.user.is_authenticated():
feedback.user = request.user
feedback.save()
submit_was_success = True
return_to_url = feedback.url or None
else:
# use GET to grab the url if set
form = FeedbackForm(initial=request.GET)
return render_to_response(
'feedback/add.html',
{
'form': form,
'submit_was_success': submit_was_success,
'return_to_url': return_to_url,
},
context_instance=RequestContext(request)
)
| Mark feedback comments starting with a html tag as spammy | ## Code Before:
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.views.decorators.csrf import csrf_protect
from models import Feedback
from forms import FeedbackForm
@csrf_protect
def add(request):
"""Gather feedback for a page, and if it is ok show a thanks message and link back to the page."""
submit_was_success = False
return_to_url = None
# If it is a post request try to create the feedback
if request.method == 'POST':
form = FeedbackForm( request.POST )
if form.is_valid():
feedback = Feedback()
feedback.url = form.cleaned_data['url']
feedback.email = form.cleaned_data['email']
feedback.comment = form.cleaned_data['comment']
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
feedback.status = 'spammy'
if request.user.is_authenticated():
feedback.user = request.user
feedback.save()
submit_was_success = True
return_to_url = feedback.url or None
else:
# use GET to grab the url if set
form = FeedbackForm(initial=request.GET)
return render_to_response(
'feedback/add.html',
{
'form': form,
'submit_was_success': submit_was_success,
'return_to_url': return_to_url,
},
context_instance=RequestContext(request)
)
## Instruction:
Mark feedback comments starting with a html tag as spammy
## Code After:
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.views.decorators.csrf import csrf_protect
from models import Feedback
from forms import FeedbackForm
import re
@csrf_protect
def add(request):
"""Gather feedback for a page, and if it is ok show a thanks message and link back to the page."""
submit_was_success = False
return_to_url = None
# If it is a post request try to create the feedback
if request.method == 'POST':
form = FeedbackForm( request.POST )
if form.is_valid():
feedback = Feedback()
feedback.url = form.cleaned_data['url']
feedback.email = form.cleaned_data['email']
feedback.comment = form.cleaned_data['comment']
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
feedback.status = 'spammy'
# if the comment starts with an html tag it is probably spam
if re.search('\A\s*<\w+>', form.cleaned_data['comment']):
feedback.status = 'spammy'
if request.user.is_authenticated():
feedback.user = request.user
feedback.save()
submit_was_success = True
return_to_url = feedback.url or None
else:
# use GET to grab the url if set
form = FeedbackForm(initial=request.GET)
return render_to_response(
'feedback/add.html',
{
'form': form,
'submit_was_success': submit_was_success,
'return_to_url': return_to_url,
},
context_instance=RequestContext(request)
)
| // ... existing code ...
from models import Feedback
from forms import FeedbackForm
import re
@csrf_protect
// ... modified code ...
# if there is any content in the honeypot field then label this comment as spammy
if form.cleaned_data['website']:
feedback.status = 'spammy'
# if the comment starts with an html tag it is probably spam
if re.search('\A\s*<\w+>', form.cleaned_data['comment']):
feedback.status = 'spammy'
// ... rest of the code ... |
c98ab8807440e3cdbb98e11c53c7f246c35614fe | dedupe/convenience.py | dedupe/convenience.py |
import collections
import dedupe.core
def dataSample(data, sample_size):
'''Randomly sample pairs of records from a data dictionary'''
random_pairs = dedupe.core.randomPairs(len(data), sample_size)
return tuple((data[k1], data[k2]) for k1, k2 in random_pairs)
def blockData(data_d, blocker):
blocks = dedupe.core.OrderedDict({})
record_blocks = dedupe.core.OrderedDict({})
key_blocks = dedupe.core.OrderedDict({})
blocker.tfIdfBlocks(data_d.iteritems())
for (record_id, record) in data_d.iteritems():
for key in blocker((record_id, record)):
blocks.setdefault(key, {}).update({record_id : record})
blocked_records = tuple(block for block in blocks.values())
return blocked_records
|
import collections
import dedupe.core
def dataSample(data, sample_size):
'''Randomly sample pairs of records from a data dictionary'''
data_list = data.values()
random_pairs = dedupe.core.randomPairs(len(data_list), sample_size)
return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs)
def blockData(data_d, blocker):
blocks = dedupe.core.OrderedDict({})
record_blocks = dedupe.core.OrderedDict({})
key_blocks = dedupe.core.OrderedDict({})
blocker.tfIdfBlocks(data_d.iteritems())
for (record_id, record) in data_d.iteritems():
for key in blocker((record_id, record)):
blocks.setdefault(key, {}).update({record_id : record})
blocked_records = tuple(block for block in blocks.values())
return blocked_records
| Change dataSample to generate indices of random pair using list of values | Change dataSample to generate indices of random pair using list of values
| Python | mit | nmiranda/dedupe,01-/dedupe,neozhangthe1/dedupe,neozhangthe1/dedupe,nmiranda/dedupe,davidkunio/dedupe,dedupeio/dedupe,dedupeio/dedupe-examples,datamade/dedupe,tfmorris/dedupe,tfmorris/dedupe,davidkunio/dedupe,01-/dedupe,datamade/dedupe,pombredanne/dedupe,dedupeio/dedupe,pombredanne/dedupe |
import collections
import dedupe.core
def dataSample(data, sample_size):
'''Randomly sample pairs of records from a data dictionary'''
+ data_list = data.values()
+ random_pairs = dedupe.core.randomPairs(len(data_list), sample_size)
- random_pairs = dedupe.core.randomPairs(len(data), sample_size)
-
- return tuple((data[k1], data[k2]) for k1, k2 in random_pairs)
+ return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs)
def blockData(data_d, blocker):
blocks = dedupe.core.OrderedDict({})
record_blocks = dedupe.core.OrderedDict({})
key_blocks = dedupe.core.OrderedDict({})
blocker.tfIdfBlocks(data_d.iteritems())
for (record_id, record) in data_d.iteritems():
for key in blocker((record_id, record)):
blocks.setdefault(key, {}).update({record_id : record})
blocked_records = tuple(block for block in blocks.values())
return blocked_records
| Change dataSample to generate indices of random pair using list of values | ## Code Before:
import collections
import dedupe.core
def dataSample(data, sample_size):
'''Randomly sample pairs of records from a data dictionary'''
random_pairs = dedupe.core.randomPairs(len(data), sample_size)
return tuple((data[k1], data[k2]) for k1, k2 in random_pairs)
def blockData(data_d, blocker):
blocks = dedupe.core.OrderedDict({})
record_blocks = dedupe.core.OrderedDict({})
key_blocks = dedupe.core.OrderedDict({})
blocker.tfIdfBlocks(data_d.iteritems())
for (record_id, record) in data_d.iteritems():
for key in blocker((record_id, record)):
blocks.setdefault(key, {}).update({record_id : record})
blocked_records = tuple(block for block in blocks.values())
return blocked_records
## Instruction:
Change dataSample to generate indices of random pair using list of values
## Code After:
import collections
import dedupe.core
def dataSample(data, sample_size):
'''Randomly sample pairs of records from a data dictionary'''
data_list = data.values()
random_pairs = dedupe.core.randomPairs(len(data_list), sample_size)
return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs)
def blockData(data_d, blocker):
blocks = dedupe.core.OrderedDict({})
record_blocks = dedupe.core.OrderedDict({})
key_blocks = dedupe.core.OrderedDict({})
blocker.tfIdfBlocks(data_d.iteritems())
for (record_id, record) in data_d.iteritems():
for key in blocker((record_id, record)):
blocks.setdefault(key, {}).update({record_id : record})
blocked_records = tuple(block for block in blocks.values())
return blocked_records
| // ... existing code ...
'''Randomly sample pairs of records from a data dictionary'''
data_list = data.values()
random_pairs = dedupe.core.randomPairs(len(data_list), sample_size)
return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs)
// ... rest of the code ... |
4aa6714284cb45a2747cea8e0f38e8fbcd8ec0bc | pymatgen/core/design_patterns.py | pymatgen/core/design_patterns.py |
from __future__ import division, unicode_literals
"""
This module defines some useful design patterns.
"""
__author__ = "Shyue Ping Ong"
__copyright__ = "Copyright 2011, The Materials Project"
__version__ = "1.0"
__maintainer__ = "Shyue Ping Ong"
__email__ = "[email protected]"
__status__ = "Production"
__date__ = "Sep 23, 2011"
class Enum(set):
"""
Creates an enum out of a set.
"""
def __getattr__(self, name):
if name in self:
return name
raise AttributeError
class NullFile(object):
"""A file object that is associated to /dev/null."""
def __new__(cls):
import os
return open(os.devnull, 'w')
def __init__(self):
"""no-op"""
class NullStream(object):
"""A fake stream with a no-op write.."""
def write(*args):
"""no-op"""
|
from __future__ import division, unicode_literals
"""
This module defines some useful design patterns.
"""
__author__ = "Shyue Ping Ong"
__copyright__ = "Copyright 2011, The Materials Project"
__version__ = "1.0"
__maintainer__ = "Shyue Ping Ong"
__email__ = "[email protected]"
__status__ = "Production"
__date__ = "Sep 23, 2011"
class Enum(set):
"""
Creates an enum out of a set.
"""
def __getattr__(self, name):
if name in self:
return name
raise AttributeError
| Move NullFile and NullStream to monty | Move NullFile and NullStream to monty
| Python | mit | Bismarrck/pymatgen,Bismarrck/pymatgen,sonium0/pymatgen,rousseab/pymatgen,Dioptas/pymatgen,migueldiascosta/pymatgen,yanikou19/pymatgen,ctoher/pymatgen,migueldiascosta/pymatgen,yanikou19/pymatgen,rousseab/pymatgen,sonium0/pymatgen,ctoher/pymatgen,ctoher/pymatgen,rousseab/pymatgen,sonium0/pymatgen,Bismarrck/pymatgen,migueldiascosta/pymatgen,Bismarrck/pymatgen,Bismarrck/pymatgen,yanikou19/pymatgen,Dioptas/pymatgen |
from __future__ import division, unicode_literals
"""
This module defines some useful design patterns.
"""
__author__ = "Shyue Ping Ong"
__copyright__ = "Copyright 2011, The Materials Project"
__version__ = "1.0"
__maintainer__ = "Shyue Ping Ong"
__email__ = "[email protected]"
__status__ = "Production"
__date__ = "Sep 23, 2011"
class Enum(set):
"""
Creates an enum out of a set.
"""
def __getattr__(self, name):
if name in self:
return name
raise AttributeError
- class NullFile(object):
- """A file object that is associated to /dev/null."""
- def __new__(cls):
- import os
- return open(os.devnull, 'w')
-
- def __init__(self):
- """no-op"""
-
-
- class NullStream(object):
- """A fake stream with a no-op write.."""
- def write(*args):
- """no-op"""
-
- | Move NullFile and NullStream to monty | ## Code Before:
from __future__ import division, unicode_literals
"""
This module defines some useful design patterns.
"""
__author__ = "Shyue Ping Ong"
__copyright__ = "Copyright 2011, The Materials Project"
__version__ = "1.0"
__maintainer__ = "Shyue Ping Ong"
__email__ = "[email protected]"
__status__ = "Production"
__date__ = "Sep 23, 2011"
class Enum(set):
"""
Creates an enum out of a set.
"""
def __getattr__(self, name):
if name in self:
return name
raise AttributeError
class NullFile(object):
"""A file object that is associated to /dev/null."""
def __new__(cls):
import os
return open(os.devnull, 'w')
def __init__(self):
"""no-op"""
class NullStream(object):
"""A fake stream with a no-op write.."""
def write(*args):
"""no-op"""
## Instruction:
Move NullFile and NullStream to monty
## Code After:
from __future__ import division, unicode_literals
"""
This module defines some useful design patterns.
"""
__author__ = "Shyue Ping Ong"
__copyright__ = "Copyright 2011, The Materials Project"
__version__ = "1.0"
__maintainer__ = "Shyue Ping Ong"
__email__ = "[email protected]"
__status__ = "Production"
__date__ = "Sep 23, 2011"
class Enum(set):
"""
Creates an enum out of a set.
"""
def __getattr__(self, name):
if name in self:
return name
raise AttributeError
| // ... existing code ...
raise AttributeError
// ... rest of the code ... |
5178318df905ed1a68d312adb3936e8748789b2b | tests/test_views.py | tests/test_views.py |
import json
import unittest
from mock import patch
from watchman import views
class TestWatchman(unittest.TestCase):
def setUp(self):
pass
@patch('watchman.views.check_databases')
def test_response_content_type_json(self, patched_check_databases):
patched_check_databases.return_value = []
response = views.status('')
self.assertEqual(response['Content-Type'], 'application/json')
@patch('watchman.views.check_databases')
def test_response_contains_expected_checks(self, patched_check_databases):
expected_checks = ['databases']
patched_check_databases.return_value = []
response = views.status('')
content = json.loads(response.content)
self.assertItemsEqual(expected_checks, content.keys())
def tearDown(self):
pass
|
import json
import unittest
from mock import patch
from watchman import views
class TestWatchman(unittest.TestCase):
def setUp(self):
pass
@patch('watchman.views.check_databases')
def test_response_content_type_json(self, patched_check_databases):
patched_check_databases.return_value = []
response = views.status('')
self.assertEqual(response['Content-Type'], 'application/json')
@patch('watchman.views.check_databases')
def test_response_contains_expected_checks(self, patched_check_databases):
expected_checks = ['databases']
patched_check_databases.return_value = []
response = views.status('')
content = json.loads(response.content)
self.assertItemsEqual(expected_checks, content.keys())
def test_check_database_handles_exception(self):
response = views.check_database('foo')
self.assertFalse(response['foo']['ok'])
self.assertEqual(response['foo']['error'], "The connection foo doesn't exist")
def tearDown(self):
pass
| Test exception handling in `check_database` | Test exception handling in `check_database`
| Python | bsd-3-clause | JBKahn/django-watchman,mwarkentin/django-watchman,mwarkentin/django-watchman,ulope/django-watchman,gerlachry/django-watchman,blag/django-watchman,JBKahn/django-watchman,blag/django-watchman,gerlachry/django-watchman,ulope/django-watchman |
import json
import unittest
from mock import patch
from watchman import views
class TestWatchman(unittest.TestCase):
def setUp(self):
pass
@patch('watchman.views.check_databases')
def test_response_content_type_json(self, patched_check_databases):
patched_check_databases.return_value = []
response = views.status('')
self.assertEqual(response['Content-Type'], 'application/json')
@patch('watchman.views.check_databases')
def test_response_contains_expected_checks(self, patched_check_databases):
expected_checks = ['databases']
patched_check_databases.return_value = []
response = views.status('')
content = json.loads(response.content)
self.assertItemsEqual(expected_checks, content.keys())
+ def test_check_database_handles_exception(self):
+ response = views.check_database('foo')
+ self.assertFalse(response['foo']['ok'])
+ self.assertEqual(response['foo']['error'], "The connection foo doesn't exist")
+
def tearDown(self):
pass
| Test exception handling in `check_database` | ## Code Before:
import json
import unittest
from mock import patch
from watchman import views
class TestWatchman(unittest.TestCase):
def setUp(self):
pass
@patch('watchman.views.check_databases')
def test_response_content_type_json(self, patched_check_databases):
patched_check_databases.return_value = []
response = views.status('')
self.assertEqual(response['Content-Type'], 'application/json')
@patch('watchman.views.check_databases')
def test_response_contains_expected_checks(self, patched_check_databases):
expected_checks = ['databases']
patched_check_databases.return_value = []
response = views.status('')
content = json.loads(response.content)
self.assertItemsEqual(expected_checks, content.keys())
def tearDown(self):
pass
## Instruction:
Test exception handling in `check_database`
## Code After:
import json
import unittest
from mock import patch
from watchman import views
class TestWatchman(unittest.TestCase):
def setUp(self):
pass
@patch('watchman.views.check_databases')
def test_response_content_type_json(self, patched_check_databases):
patched_check_databases.return_value = []
response = views.status('')
self.assertEqual(response['Content-Type'], 'application/json')
@patch('watchman.views.check_databases')
def test_response_contains_expected_checks(self, patched_check_databases):
expected_checks = ['databases']
patched_check_databases.return_value = []
response = views.status('')
content = json.loads(response.content)
self.assertItemsEqual(expected_checks, content.keys())
def test_check_database_handles_exception(self):
response = views.check_database('foo')
self.assertFalse(response['foo']['ok'])
self.assertEqual(response['foo']['error'], "The connection foo doesn't exist")
def tearDown(self):
pass
| // ... existing code ...
self.assertItemsEqual(expected_checks, content.keys())
def test_check_database_handles_exception(self):
response = views.check_database('foo')
self.assertFalse(response['foo']['ok'])
self.assertEqual(response['foo']['error'], "The connection foo doesn't exist")
def tearDown(self):
pass
// ... rest of the code ... |
80e4caad24bceabd8e15133a96a6aaddd9a97c07 | code/type_null_true_false.py | code/type_null_true_false.py | def if_value(values):
print('"if value":')
for k, v in values:
print("%s - %s" % (k, 'true' if v else 'false'))
print()
def nil_value(values):
print('"if value is None":')
for k, v in values:
print("%s - %s" % (k, 'true' if v is None else 'false'))
print()
def empty_value(values):
print('"if len(value)":')
for k, v in values:
try:
print("%s - %s" % (k, 'true' if len(v) else 'false'))
except TypeError as e:
print("%s - %s" % (k, e))
values = [
("'string'", 'string'),
("''", ''),
('[1, 2, 3]', [1, 2, 3]),
('[]', []),
('5', 5),
('0', 0),
(True, True),
(False, False),
(None, None),
]
if_value(values)
nil_value(values)
empty_value(values)
| def check(label, fn, values):
print(label)
for value in values:
try:
result = 'true' if fn(value) else 'false'
except TypeError as e:
result = 'error: %s' % e
print(" %-9r - %s" % (value, result))
print()
values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None]
check('if value:', lambda v: v, values)
check('if value is None:', lambda v: v is None, values)
check('if len(value):', lambda v: len(v), values)
| Refactor Null/True/False to look more pythonic | Refactor Null/True/False to look more pythonic
| Python | mit | evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,Evmorov/ruby-coffeescript,evmorov/lang-compare,evmorov/lang-compare | - def if_value(values):
- print('"if value":')
+ def check(label, fn, values):
+ print(label)
- for k, v in values:
+ for value in values:
- print("%s - %s" % (k, 'true' if v else 'false'))
+ try:
+ result = 'true' if fn(value) else 'false'
+ except TypeError as e:
+ result = 'error: %s' % e
+ print(" %-9r - %s" % (value, result))
- print()
+ print()
+ values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None]
- def nil_value(values):
- print('"if value is None":')
- for k, v in values:
- print("%s - %s" % (k, 'true' if v is None else 'false'))
- print()
+ check('if value:', lambda v: v, values)
+ check('if value is None:', lambda v: v is None, values)
+ check('if len(value):', lambda v: len(v), values)
- def empty_value(values):
- print('"if len(value)":')
- for k, v in values:
- try:
- print("%s - %s" % (k, 'true' if len(v) else 'false'))
- except TypeError as e:
- print("%s - %s" % (k, e))
- values = [
- ("'string'", 'string'),
- ("''", ''),
- ('[1, 2, 3]', [1, 2, 3]),
- ('[]', []),
- ('5', 5),
- ('0', 0),
- (True, True),
- (False, False),
- (None, None),
- ]
-
- if_value(values)
- nil_value(values)
- empty_value(values)
- | Refactor Null/True/False to look more pythonic | ## Code Before:
def if_value(values):
print('"if value":')
for k, v in values:
print("%s - %s" % (k, 'true' if v else 'false'))
print()
def nil_value(values):
print('"if value is None":')
for k, v in values:
print("%s - %s" % (k, 'true' if v is None else 'false'))
print()
def empty_value(values):
print('"if len(value)":')
for k, v in values:
try:
print("%s - %s" % (k, 'true' if len(v) else 'false'))
except TypeError as e:
print("%s - %s" % (k, e))
values = [
("'string'", 'string'),
("''", ''),
('[1, 2, 3]', [1, 2, 3]),
('[]', []),
('5', 5),
('0', 0),
(True, True),
(False, False),
(None, None),
]
if_value(values)
nil_value(values)
empty_value(values)
## Instruction:
Refactor Null/True/False to look more pythonic
## Code After:
def check(label, fn, values):
print(label)
for value in values:
try:
result = 'true' if fn(value) else 'false'
except TypeError as e:
result = 'error: %s' % e
print(" %-9r - %s" % (value, result))
print()
values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None]
check('if value:', lambda v: v, values)
check('if value is None:', lambda v: v is None, values)
check('if len(value):', lambda v: len(v), values)
| ...
def check(label, fn, values):
print(label)
for value in values:
try:
result = 'true' if fn(value) else 'false'
except TypeError as e:
result = 'error: %s' % e
print(" %-9r - %s" % (value, result))
print()
values = ['string', '', [1, 2, 3], [], 5, 0, True, False, None]
check('if value:', lambda v: v, values)
check('if value is None:', lambda v: v is None, values)
check('if len(value):', lambda v: len(v), values)
... |
04a7de877c50bc84428e7bb7d30b1c6cac00a59f | ipywidgets/widgets/tests/test_widget_selection.py | ipywidgets/widgets/tests/test_widget_selection.py |
import warnings
from unittest import TestCase
from ipywidgets import Dropdown
class TestDropdown(TestCase):
def test_construction(self):
Dropdown()
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("module")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
assert len(w) == 1
assert issubclass(w[-1].category, DeprecationWarning)
assert "deprecated" in str(w[-1].message)
|
import warnings
from unittest import TestCase
from ipywidgets import Dropdown
class TestDropdown(TestCase):
def test_construction(self):
Dropdown()
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("always")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
assert len(w) > 0
assert issubclass(w[-1].category, DeprecationWarning)
assert "Support for mapping types has been deprecated" in str(w[-1].message)
| Use simplefilter('always') for testing the warning | Use simplefilter('always') for testing the warning
* Use `warnings.simplefilter('always')` for DeprecationWarning
* More specific test on warning message
| Python | bsd-3-clause | jupyter-widgets/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets,jupyter-widgets/ipywidgets,ipython/ipywidgets,jupyter-widgets/ipywidgets,ipython/ipywidgets,jupyter-widgets/ipywidgets,SylvainCorlay/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets,SylvainCorlay/ipywidgets,ipython/ipywidgets |
import warnings
from unittest import TestCase
from ipywidgets import Dropdown
class TestDropdown(TestCase):
def test_construction(self):
Dropdown()
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
- warnings.simplefilter("module")
+ warnings.simplefilter("always")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
- assert len(w) == 1
+ assert len(w) > 0
assert issubclass(w[-1].category, DeprecationWarning)
- assert "deprecated" in str(w[-1].message)
+ assert "Support for mapping types has been deprecated" in str(w[-1].message)
| Use simplefilter('always') for testing the warning | ## Code Before:
import warnings
from unittest import TestCase
from ipywidgets import Dropdown
class TestDropdown(TestCase):
def test_construction(self):
Dropdown()
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("module")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
assert len(w) == 1
assert issubclass(w[-1].category, DeprecationWarning)
assert "deprecated" in str(w[-1].message)
## Instruction:
Use simplefilter('always') for testing the warning
## Code After:
import warnings
from unittest import TestCase
from ipywidgets import Dropdown
class TestDropdown(TestCase):
def test_construction(self):
Dropdown()
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("always")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
assert len(w) > 0
assert issubclass(w[-1].category, DeprecationWarning)
assert "Support for mapping types has been deprecated" in str(w[-1].message)
| // ... existing code ...
def test_deprecation_warning_mapping_options(self):
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("always")
Dropdown(options={'One': 1, 'Two': 2, 'Three': 3})
assert len(w) > 0
assert issubclass(w[-1].category, DeprecationWarning)
assert "Support for mapping types has been deprecated" in str(w[-1].message)
// ... rest of the code ... |
4425aa1170a1acd3ed69c32ba5e3885301593524 | salt/returners/redis_return.py | salt/returners/redis_return.py | '''
Return data to a redis server
To enable this returner the minion will need the python client for redis
installed and the following values configured in the minion or master
config, these are the defaults:
redis.db: '0'
redis.host: 'salt'
redis.port: 6379
'''
# Import python libs
import json
try:
import redis
has_redis = True
except ImportError:
has_redis = False
def __virtual__():
if not has_redis:
return False
return 'redis_return'
def _get_serv():
'''
Return a redis server object
'''
return redis.Redis(
host=__salt__['config.option']('redis.host'),
port=__salt__['config.option']('redis.port'),
db=__salt__['config.option']('redis.db'))
def returner(ret):
'''
Return data to a redis data store
'''
serv = _get_serv()
serv.sadd('{0}:jobs'.format(ret['id']))
serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return'])))
serv.sadd('jobs', ret['jid'])
serv.sadd(ret['jid'], ret['id'])
| '''
Return data to a redis server
To enable this returner the minion will need the python client for redis
installed and the following values configured in the minion or master
config, these are the defaults:
redis.db: '0'
redis.host: 'salt'
redis.port: 6379
'''
# Import python libs
import json
try:
import redis
has_redis = True
except ImportError:
has_redis = False
def __virtual__():
if not has_redis:
return False
return 'redis_return'
def _get_serv():
'''
Return a redis server object
'''
return redis.Redis(
host=__salt__['config.option']('redis.host'),
port=__salt__['config.option']('redis.port'),
db=__salt__['config.option']('redis.db'))
def returner(ret):
'''
Return data to a redis data store
'''
serv = _get_serv()
serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret))
serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid'])
serv.sadd('minions', ret['id'])
| Restructure redis returner, since it did notwork before anyway | Restructure redis returner, since it did notwork before anyway
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | '''
Return data to a redis server
To enable this returner the minion will need the python client for redis
installed and the following values configured in the minion or master
config, these are the defaults:
redis.db: '0'
redis.host: 'salt'
redis.port: 6379
'''
# Import python libs
import json
try:
import redis
has_redis = True
except ImportError:
has_redis = False
def __virtual__():
if not has_redis:
return False
return 'redis_return'
def _get_serv():
'''
Return a redis server object
'''
return redis.Redis(
host=__salt__['config.option']('redis.host'),
port=__salt__['config.option']('redis.port'),
db=__salt__['config.option']('redis.db'))
def returner(ret):
'''
Return data to a redis data store
'''
serv = _get_serv()
- serv.sadd('{0}:jobs'.format(ret['id']))
- serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return'])))
+ serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret))
+ serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid'])
- serv.sadd('jobs', ret['jid'])
+ serv.sadd('minions', ret['id'])
- serv.sadd(ret['jid'], ret['id'])
| Restructure redis returner, since it did notwork before anyway | ## Code Before:
'''
Return data to a redis server
To enable this returner the minion will need the python client for redis
installed and the following values configured in the minion or master
config, these are the defaults:
redis.db: '0'
redis.host: 'salt'
redis.port: 6379
'''
# Import python libs
import json
try:
import redis
has_redis = True
except ImportError:
has_redis = False
def __virtual__():
if not has_redis:
return False
return 'redis_return'
def _get_serv():
'''
Return a redis server object
'''
return redis.Redis(
host=__salt__['config.option']('redis.host'),
port=__salt__['config.option']('redis.port'),
db=__salt__['config.option']('redis.db'))
def returner(ret):
'''
Return data to a redis data store
'''
serv = _get_serv()
serv.sadd('{0}:jobs'.format(ret['id']))
serv.set('{0}:{1}'.format(ret['jid'], json.dumps(ret['return'])))
serv.sadd('jobs', ret['jid'])
serv.sadd(ret['jid'], ret['id'])
## Instruction:
Restructure redis returner, since it did notwork before anyway
## Code After:
'''
Return data to a redis server
To enable this returner the minion will need the python client for redis
installed and the following values configured in the minion or master
config, these are the defaults:
redis.db: '0'
redis.host: 'salt'
redis.port: 6379
'''
# Import python libs
import json
try:
import redis
has_redis = True
except ImportError:
has_redis = False
def __virtual__():
if not has_redis:
return False
return 'redis_return'
def _get_serv():
'''
Return a redis server object
'''
return redis.Redis(
host=__salt__['config.option']('redis.host'),
port=__salt__['config.option']('redis.port'),
db=__salt__['config.option']('redis.db'))
def returner(ret):
'''
Return data to a redis data store
'''
serv = _get_serv()
serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret))
serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid'])
serv.sadd('minions', ret['id'])
| // ... existing code ...
'''
serv = _get_serv()
serv.set('{0}:{1}'.format(ret['id'], ret['jid']), json.dumps(ret))
serv.lpush('{0}:{1}'.format(ret['id'], ret['fun']), ret['jid'])
serv.sadd('minions', ret['id'])
// ... rest of the code ... |
6a4046aafe43930c202e2f18a55b1cd8517d95f9 | testanalyzer/javaanalyzer.py | testanalyzer/javaanalyzer.py | import re
from fileanalyzer import FileAnalyzer
class JavaAnalyzer(FileAnalyzer):
def get_class_count(self, content):
return len(
re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content))
# TODO: Accept angle brackets and decline "else if"
def get_function_count(self, content):
return len(
re.findall(
"[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{",
content))
| import re
from fileanalyzer import FileAnalyzer
class JavaAnalyzer(FileAnalyzer):
def get_class_count(self, content):
return len(
re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content))
def get_function_count(self, content):
matches = re.findall(
"[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{",
content)
matches = [
m for m in matches
if "if " not in m.strip() and "if(" not in m.strip()
]
return len(matches)
| Fix regex to match generics | Fix regex to match generics
| Python | mpl-2.0 | CheriPai/TestAnalyzer,CheriPai/TestAnalyzer,CheriPai/TestAnalyzer | import re
from fileanalyzer import FileAnalyzer
class JavaAnalyzer(FileAnalyzer):
def get_class_count(self, content):
return len(
- re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content))
+ re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content))
- # TODO: Accept angle brackets and decline "else if"
def get_function_count(self, content):
- return len(
- re.findall(
+ matches = re.findall(
- "[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{",
+ "[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{",
- content))
+ content)
+ matches = [
+ m for m in matches
+ if "if " not in m.strip() and "if(" not in m.strip()
+ ]
+ return len(matches)
| Fix regex to match generics | ## Code Before:
import re
from fileanalyzer import FileAnalyzer
class JavaAnalyzer(FileAnalyzer):
def get_class_count(self, content):
return len(
re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_]+ *\n*\{", content))
# TODO: Accept angle brackets and decline "else if"
def get_function_count(self, content):
return len(
re.findall(
"[a-zA-Z ]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\] \n]*\)[a-zA-Z \n]*\{",
content))
## Instruction:
Fix regex to match generics
## Code After:
import re
from fileanalyzer import FileAnalyzer
class JavaAnalyzer(FileAnalyzer):
def get_class_count(self, content):
return len(
re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content))
def get_function_count(self, content):
matches = re.findall(
"[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{",
content)
matches = [
m for m in matches
if "if " not in m.strip() and "if(" not in m.strip()
]
return len(matches)
| # ... existing code ...
def get_class_count(self, content):
return len(
re.findall("[a-zA-Z ]*class +[a-zA-Z0-9_<>, ]+\n*\{", content))
def get_function_count(self, content):
matches = re.findall(
"[a-zA-Z <>]+ +[a-zA-Z0-9_]+ *\n*\([a-zA-Z0-9_,\[\]<>\?\. \n]*\)[a-zA-Z \n]*\{",
content)
matches = [
m for m in matches
if "if " not in m.strip() and "if(" not in m.strip()
]
return len(matches)
# ... rest of the code ... |
133617660fe96a817b47d4d0fba4cfa7567dcafb | exceptional.py | exceptional.py | """A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1])) | """A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | Use two return statements and remove printing | Use two return statements and remove printing
| Python | mit | kentoj/python-fundamentals | """A module to demonstrate exceptions."""
import sys
def convert(item):
- '''
+ """
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
- '''
+ """
try:
- x = int(item)
+ return int(item)
- print(str.format('Conversion succeeded! x= {}', x))
- except ValueError:
+ except (ValueError, TypeError):
- print('Conversion Failed')
- x = -1
- return x
+ return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | Use two return statements and remove printing | ## Code Before:
"""A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1]))
## Instruction:
Use two return statements and remove printing
## Code After:
"""A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | ...
def convert(item):
"""
Convert to an integer.
Args:
...
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
... |
8608283592338960c80113ff4d68f42936ddb969 | linter.py | linter.py |
"""This module exports the Perl plugin class."""
import shlex
from SublimeLinter.lint import Linter, util
class Perl(Linter):
"""Provides an interface to perl -c."""
syntax = ('modernperl', 'perl')
executable = 'perl'
base_cmd = ('perl -c')
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
def cmd(self):
"""
Return the command line to execute.
Overridden so we can add include paths based on the 'include_dirs'
settings.
"""
full_cmd = self.base_cmd
settings = self.get_view_settings()
include_dirs = settings.get('include_dirs', [])
if include_dirs:
full_cmd += ' ' . join([' -I ' + shlex.quote(include)
for include in include_dirs])
return full_cmd
|
"""This module exports the Perl plugin class."""
import shlex
from SublimeLinter.lint import Linter, util
class Perl(Linter):
"""Provides an interface to perl -c."""
syntax = ('modernperl', 'perl')
executable = 'perl'
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
def cmd(self):
"""
Return the command line to execute.
Overridden so we can add include paths based on the 'include_dirs'
settings.
"""
command = [self.executable_path, '-c']
include_dirs = self.get_view_settings().get('include_dirs', [])
for e in include_dirs:
command.append('-I')
command.append(shlex.quote(e))
return command
| Clean up include dir code | Clean up include dir code
| Python | mit | oschwald/SublimeLinter-perl |
"""This module exports the Perl plugin class."""
import shlex
from SublimeLinter.lint import Linter, util
class Perl(Linter):
"""Provides an interface to perl -c."""
syntax = ('modernperl', 'perl')
executable = 'perl'
- base_cmd = ('perl -c')
+
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
def cmd(self):
"""
Return the command line to execute.
Overridden so we can add include paths based on the 'include_dirs'
settings.
"""
- full_cmd = self.base_cmd
+ command = [self.executable_path, '-c']
- settings = self.get_view_settings()
+ include_dirs = self.get_view_settings().get('include_dirs', [])
- include_dirs = settings.get('include_dirs', [])
+ for e in include_dirs:
+ command.append('-I')
+ command.append(shlex.quote(e))
+ return command
- if include_dirs:
- full_cmd += ' ' . join([' -I ' + shlex.quote(include)
- for include in include_dirs])
- return full_cmd
- | Clean up include dir code | ## Code Before:
"""This module exports the Perl plugin class."""
import shlex
from SublimeLinter.lint import Linter, util
class Perl(Linter):
"""Provides an interface to perl -c."""
syntax = ('modernperl', 'perl')
executable = 'perl'
base_cmd = ('perl -c')
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
def cmd(self):
"""
Return the command line to execute.
Overridden so we can add include paths based on the 'include_dirs'
settings.
"""
full_cmd = self.base_cmd
settings = self.get_view_settings()
include_dirs = settings.get('include_dirs', [])
if include_dirs:
full_cmd += ' ' . join([' -I ' + shlex.quote(include)
for include in include_dirs])
return full_cmd
## Instruction:
Clean up include dir code
## Code After:
"""This module exports the Perl plugin class."""
import shlex
from SublimeLinter.lint import Linter, util
class Perl(Linter):
"""Provides an interface to perl -c."""
syntax = ('modernperl', 'perl')
executable = 'perl'
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
def cmd(self):
"""
Return the command line to execute.
Overridden so we can add include paths based on the 'include_dirs'
settings.
"""
command = [self.executable_path, '-c']
include_dirs = self.get_view_settings().get('include_dirs', [])
for e in include_dirs:
command.append('-I')
command.append(shlex.quote(e))
return command
| ...
syntax = ('modernperl', 'perl')
executable = 'perl'
regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?'
error_stream = util.STREAM_STDERR
...
"""
command = [self.executable_path, '-c']
include_dirs = self.get_view_settings().get('include_dirs', [])
for e in include_dirs:
command.append('-I')
command.append(shlex.quote(e))
return command
... |
403f23ae486c14066e0a93c7deca91c5fbc15b87 | plugins/brian.py | plugins/brian.py | """Displays a randomly generated witticism from Brian Chu himself."""
import json
import random
__match__ = r"!brian"
with open('plugins/brian_corpus/cache.json', 'r') as infile:
cache = json.load(infile)
with open('plugins/brian_corpus/phrases.json', 'r') as infile:
phrases = json.load(infile)
def generate_phrase(phrases, cache):
seed_phrase = []
while len(seed_phrase) < 3:
seed_phrase = random.choice(phrases).split()
w1, w2 = seed_phrase[:2]
chosen = [w1, w2]
while "{}|{}".format(w1, w2) in cache:
choice = random.choice(cache["{}|{}".format(w1, w2)])
w1, w2 = w2, choice
chosen.append(choice)
return ' '.join(chosen)
def on_message(bot, channel, user, message):
return '> {} ~brian'.format(generate_phrase(phrases, cache))
| """Displays a randomly generated witticism from Brian Chu himself."""
import json
import random
__match__ = r"!brian"
attribution = [
"salad master",
"esquire",
"the one and only",
"startup enthusiast",
"boba king",
"not-dictator",
"normal citizen",
"ping-pong expert"
]
with open('plugins/brian_corpus/phrases.json', 'r') as infile:
phrases = json.load(infile)
with open('plugins/brian_corpus/cache.json', 'r') as infile:
cache = json.load(infile)
def generate_phrase(phrases, cache, max_length=40):
seed_phrase = []
while len(seed_phrase) < 2:
seed_phrase = random.choice(phrases).split()
w1, = seed_phrase[:1]
chosen = [w1]
while w1 in cache and len(chosen)<max_length:
w1 = random.choice(cache[w1])
chosen.append(w1)
return ' '.join(chosen)
def on_message(bot, channel, user, message):
return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache),
random.choice(attribution))
| Use bigrams in Markov chain generator | Use bigrams in Markov chain generator
| Python | mit | kvchen/keffbot,kvchen/keffbot-py | """Displays a randomly generated witticism from Brian Chu himself."""
import json
import random
__match__ = r"!brian"
+ attribution = [
+ "salad master",
+ "esquire",
+ "the one and only",
+ "startup enthusiast",
+ "boba king",
+ "not-dictator",
+ "normal citizen",
+ "ping-pong expert"
+ ]
+
+
+ with open('plugins/brian_corpus/phrases.json', 'r') as infile:
+ phrases = json.load(infile)
with open('plugins/brian_corpus/cache.json', 'r') as infile:
cache = json.load(infile)
- with open('plugins/brian_corpus/phrases.json', 'r') as infile:
- phrases = json.load(infile)
-
- def generate_phrase(phrases, cache):
+ def generate_phrase(phrases, cache, max_length=40):
seed_phrase = []
- while len(seed_phrase) < 3:
+ while len(seed_phrase) < 2:
seed_phrase = random.choice(phrases).split()
- w1, w2 = seed_phrase[:2]
+ w1, = seed_phrase[:1]
- chosen = [w1, w2]
+ chosen = [w1]
+ while w1 in cache and len(chosen)<max_length:
+ w1 = random.choice(cache[w1])
- while "{}|{}".format(w1, w2) in cache:
- choice = random.choice(cache["{}|{}".format(w1, w2)])
- w1, w2 = w2, choice
- chosen.append(choice)
+ chosen.append(w1)
return ' '.join(chosen)
def on_message(bot, channel, user, message):
- return '> {} ~brian'.format(generate_phrase(phrases, cache))
+ return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache),
+ random.choice(attribution))
| Use bigrams in Markov chain generator | ## Code Before:
"""Displays a randomly generated witticism from Brian Chu himself."""
import json
import random
__match__ = r"!brian"
with open('plugins/brian_corpus/cache.json', 'r') as infile:
cache = json.load(infile)
with open('plugins/brian_corpus/phrases.json', 'r') as infile:
phrases = json.load(infile)
def generate_phrase(phrases, cache):
seed_phrase = []
while len(seed_phrase) < 3:
seed_phrase = random.choice(phrases).split()
w1, w2 = seed_phrase[:2]
chosen = [w1, w2]
while "{}|{}".format(w1, w2) in cache:
choice = random.choice(cache["{}|{}".format(w1, w2)])
w1, w2 = w2, choice
chosen.append(choice)
return ' '.join(chosen)
def on_message(bot, channel, user, message):
return '> {} ~brian'.format(generate_phrase(phrases, cache))
## Instruction:
Use bigrams in Markov chain generator
## Code After:
"""Displays a randomly generated witticism from Brian Chu himself."""
import json
import random
__match__ = r"!brian"
attribution = [
"salad master",
"esquire",
"the one and only",
"startup enthusiast",
"boba king",
"not-dictator",
"normal citizen",
"ping-pong expert"
]
with open('plugins/brian_corpus/phrases.json', 'r') as infile:
phrases = json.load(infile)
with open('plugins/brian_corpus/cache.json', 'r') as infile:
cache = json.load(infile)
def generate_phrase(phrases, cache, max_length=40):
seed_phrase = []
while len(seed_phrase) < 2:
seed_phrase = random.choice(phrases).split()
w1, = seed_phrase[:1]
chosen = [w1]
while w1 in cache and len(chosen)<max_length:
w1 = random.choice(cache[w1])
chosen.append(w1)
return ' '.join(chosen)
def on_message(bot, channel, user, message):
return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache),
random.choice(attribution))
| ...
__match__ = r"!brian"
attribution = [
"salad master",
"esquire",
"the one and only",
"startup enthusiast",
"boba king",
"not-dictator",
"normal citizen",
"ping-pong expert"
]
with open('plugins/brian_corpus/phrases.json', 'r') as infile:
phrases = json.load(infile)
with open('plugins/brian_corpus/cache.json', 'r') as infile:
...
cache = json.load(infile)
def generate_phrase(phrases, cache, max_length=40):
seed_phrase = []
while len(seed_phrase) < 2:
seed_phrase = random.choice(phrases).split()
w1, = seed_phrase[:1]
chosen = [w1]
while w1 in cache and len(chosen)<max_length:
w1 = random.choice(cache[w1])
chosen.append(w1)
return ' '.join(chosen)
...
def on_message(bot, channel, user, message):
return '> {} ~ Brian Chu, {}'.format(generate_phrase(phrases, cache),
random.choice(attribution))
... |
16b3dc1f8c762a751e1476d679391f3bbc82cd5d | python-prefix.py | python-prefix.py |
import sys
import os.path
import site
def main():
'''\
Check if the given prefix is included in sys.path for the given
python version; if not find an alternate valid prefix. Print the
result to standard out.
'''
if len(sys.argv) != 3:
msg = 'usage: %s <prefix> <python version>\n' % \
os.path.basename(sys.argv[0])
sys.stderr.write(msg)
return 1
python_prefix = sys.argv[1]
python_version = sys.argv[2]
path = '%s/lib/python%s' % (python_prefix, python_version)
path = os.path.normpath(path)
if path[-1] != '/':
path = path + '/'
prefix = None
for p in sys.path:
if p.startswith(path):
prefix = path
break
if not prefix:
prefix = site.PREFIXES[-1]
sys.stdout.write('%s\n' % prefix)
return 0
if __name__ == '__main__':
sys.exit(main())
|
import sys
import os.path
import site
def main():
'''\
Check if the given prefix is included in sys.path for the given
python version; if not find an alternate valid prefix. Print the
result to standard out.
'''
if len(sys.argv) != 3:
msg = 'usage: %s <prefix> <python version>\n' % \
os.path.basename(sys.argv[0])
sys.stderr.write(msg)
return 1
python_prefix = sys.argv[1]
python_version = sys.argv[2]
path = '%s/lib/python%s' % (python_prefix, python_version)
path = os.path.normpath(path)
if path[-1] != '/':
path = path + '/'
prefix = None
for p in sys.path:
if p.startswith(path):
prefix = python_prefix
break
if not prefix:
prefix = site.PREFIXES[-1]
sys.stdout.write('%s\n' % prefix)
return 0
if __name__ == '__main__':
sys.exit(main())
| Fix typo in previous commit. | Fix typo in previous commit. | Python | bsd-2-clause | marek-sezemsky/coreemu,tectronics/coreemu,marek-sezemsky/coreemu,guidotack/coreemu,guidotack/coreemu,tectronics/coreemu,tectronics/coreemu,gregtampa/coreemu,guidotack/coreemu,gregtampa/coreemu,marek-sezemsky/coreemu,gregtampa/coreemu |
import sys
import os.path
import site
def main():
'''\
Check if the given prefix is included in sys.path for the given
python version; if not find an alternate valid prefix. Print the
result to standard out.
'''
if len(sys.argv) != 3:
msg = 'usage: %s <prefix> <python version>\n' % \
os.path.basename(sys.argv[0])
sys.stderr.write(msg)
return 1
python_prefix = sys.argv[1]
python_version = sys.argv[2]
path = '%s/lib/python%s' % (python_prefix, python_version)
path = os.path.normpath(path)
if path[-1] != '/':
path = path + '/'
prefix = None
for p in sys.path:
if p.startswith(path):
- prefix = path
+ prefix = python_prefix
break
if not prefix:
prefix = site.PREFIXES[-1]
sys.stdout.write('%s\n' % prefix)
return 0
if __name__ == '__main__':
sys.exit(main())
| Fix typo in previous commit. | ## Code Before:
import sys
import os.path
import site
def main():
'''\
Check if the given prefix is included in sys.path for the given
python version; if not find an alternate valid prefix. Print the
result to standard out.
'''
if len(sys.argv) != 3:
msg = 'usage: %s <prefix> <python version>\n' % \
os.path.basename(sys.argv[0])
sys.stderr.write(msg)
return 1
python_prefix = sys.argv[1]
python_version = sys.argv[2]
path = '%s/lib/python%s' % (python_prefix, python_version)
path = os.path.normpath(path)
if path[-1] != '/':
path = path + '/'
prefix = None
for p in sys.path:
if p.startswith(path):
prefix = path
break
if not prefix:
prefix = site.PREFIXES[-1]
sys.stdout.write('%s\n' % prefix)
return 0
if __name__ == '__main__':
sys.exit(main())
## Instruction:
Fix typo in previous commit.
## Code After:
import sys
import os.path
import site
def main():
'''\
Check if the given prefix is included in sys.path for the given
python version; if not find an alternate valid prefix. Print the
result to standard out.
'''
if len(sys.argv) != 3:
msg = 'usage: %s <prefix> <python version>\n' % \
os.path.basename(sys.argv[0])
sys.stderr.write(msg)
return 1
python_prefix = sys.argv[1]
python_version = sys.argv[2]
path = '%s/lib/python%s' % (python_prefix, python_version)
path = os.path.normpath(path)
if path[-1] != '/':
path = path + '/'
prefix = None
for p in sys.path:
if p.startswith(path):
prefix = python_prefix
break
if not prefix:
prefix = site.PREFIXES[-1]
sys.stdout.write('%s\n' % prefix)
return 0
if __name__ == '__main__':
sys.exit(main())
| ...
for p in sys.path:
if p.startswith(path):
prefix = python_prefix
break
if not prefix:
... |
524ee1cd2f56f6fe968f409d37cbd2af1621e7f3 | framework/guid/model.py | framework/guid/model.py | from framework import StoredObject, fields
class Guid(StoredObject):
_id = fields.StringField()
referent = fields.AbstractForeignField()
_meta = {
'optimistic': True,
}
class GuidStoredObject(StoredObject):
# Redirect to content using URL redirect by default
redirect_mode = 'redirect'
def _ensure_guid(self):
"""Create GUID record if current record doesn't already have one, then
point GUID to self.
"""
# Create GUID with specified ID if ID provided
if self._primary_key:
# Done if GUID already exists
guid = Guid.load(self._primary_key)
if guid is not None:
return
# Create GUID
guid = Guid(
_id=self._primary_key,
referent=self
)
guid.save()
# Else create GUID optimistically
else:
# Create GUID
guid = Guid()
guid.save()
guid.referent = (guid._primary_key, self._name)
guid.save()
# Set primary key to GUID key
self._primary_key = guid._primary_key
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
rv = super(GuidStoredObject, self).save(*args, **kwargs)
self._ensure_guid()
return rv
@property
def annotations(self):
""" Get meta-data annotations associated with object. """
return self.metadata__annotated
| from framework import StoredObject, fields
class Guid(StoredObject):
_id = fields.StringField()
referent = fields.AbstractForeignField()
_meta = {
'optimistic': True,
}
class GuidStoredObject(StoredObject):
# Redirect to content using URL redirect by default
redirect_mode = 'redirect'
def _ensure_guid(self):
"""Create GUID record if current record doesn't already have one, then
point GUID to self.
"""
# Create GUID with specified ID if ID provided
if self._primary_key:
# Done if GUID already exists
guid = Guid.load(self._primary_key)
if guid is not None:
return
# Create GUID
guid = Guid(
_id=self._primary_key,
referent=self
)
guid.save()
# Else create GUID optimistically
else:
# Create GUID
guid = Guid()
guid.save()
guid.referent = (guid._primary_key, self._name)
guid.save()
# Set primary key to GUID key
self._primary_key = guid._primary_key
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
self._ensure_guid()
return super(GuidStoredObject, self).save(*args, **kwargs)
@property
def annotations(self):
""" Get meta-data annotations associated with object. """
return self.metadata__annotated
| Fix last commit: Must ensure GUID before saving so that PK is defined | Fix last commit: Must ensure GUID before saving so that PK is defined
| Python | apache-2.0 | zkraime/osf.io,emetsger/osf.io,RomanZWang/osf.io,chennan47/osf.io,TomHeatwole/osf.io,adlius/osf.io,cwisecarver/osf.io,petermalcolm/osf.io,mfraezz/osf.io,wearpants/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,samanehsan/osf.io,CenterForOpenScience/osf.io,caseyrollins/osf.io,felliott/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,jmcarp/osf.io,jeffreyliu3230/osf.io,sbt9uc/osf.io,Ghalko/osf.io,cwisecarver/osf.io,aaxelb/osf.io,pattisdr/osf.io,SSJohns/osf.io,billyhunt/osf.io,danielneis/osf.io,GageGaskins/osf.io,ZobairAlijan/osf.io,ckc6cz/osf.io,acshi/osf.io,cslzchen/osf.io,njantrania/osf.io,samchrisinger/osf.io,acshi/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,lamdnhan/osf.io,njantrania/osf.io,billyhunt/osf.io,zamattiac/osf.io,felliott/osf.io,rdhyee/osf.io,alexschiller/osf.io,mluo613/osf.io,zachjanicki/osf.io,Nesiehr/osf.io,alexschiller/osf.io,crcresearch/osf.io,kwierman/osf.io,arpitar/osf.io,cwisecarver/osf.io,TomHeatwole/osf.io,abought/osf.io,mfraezz/osf.io,ckc6cz/osf.io,MerlinZhang/osf.io,fabianvf/osf.io,himanshuo/osf.io,mattclark/osf.io,mluo613/osf.io,asanfilippo7/osf.io,Johnetordoff/osf.io,samchrisinger/osf.io,TomBaxter/osf.io,himanshuo/osf.io,mluo613/osf.io,zkraime/osf.io,KAsante95/osf.io,revanthkolli/osf.io,himanshuo/osf.io,barbour-em/osf.io,adlius/osf.io,caneruguz/osf.io,fabianvf/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,binoculars/osf.io,doublebits/osf.io,ckc6cz/osf.io,caseyrygt/osf.io,pattisdr/osf.io,barbour-em/osf.io,laurenrevere/osf.io,revanthkolli/osf.io,jeffreyliu3230/osf.io,barbour-em/osf.io,cldershem/osf.io,monikagrabowska/osf.io,petermalcolm/osf.io,danielneis/osf.io,jinluyuan/osf.io,danielneis/osf.io,chrisseto/osf.io,lyndsysimon/osf.io,jeffreyliu3230/osf.io,jinluyuan/osf.io,jeffreyliu3230/osf.io,chennan47/osf.io,ticklemepierce/osf.io,aaxelb/osf.io,mluke93/osf.io,samchrisinger/osf.io,SSJohns/osf.io,binoculars/osf.io,reinaH/osf.io,mluo613/osf.io,cosenal/osf.io,CenterForOpenScience/osf.io,Ghalko/osf.io,caneruguz/osf.io,brandonPurvis/osf.io,GageGaskins/osf.io,DanielSBrown/osf.io,mfraezz/osf.io,bdyetton/prettychart,jinluyuan/osf.io,baylee-d/osf.io,icereval/osf.io,kushG/osf.io,reinaH/osf.io,jolene-esposito/osf.io,mattclark/osf.io,ckc6cz/osf.io,reinaH/osf.io,mattclark/osf.io,zachjanicki/osf.io,wearpants/osf.io,asanfilippo7/osf.io,cldershem/osf.io,cldershem/osf.io,laurenrevere/osf.io,lyndsysimon/osf.io,cslzchen/osf.io,zkraime/osf.io,GageGaskins/osf.io,brianjgeiger/osf.io,mluke93/osf.io,zamattiac/osf.io,GaryKriebel/osf.io,TomBaxter/osf.io,lamdnhan/osf.io,brianjgeiger/osf.io,dplorimer/osf,cosenal/osf.io,aaxelb/osf.io,erinspace/osf.io,kushG/osf.io,saradbowman/osf.io,chrisseto/osf.io,kushG/osf.io,asanfilippo7/osf.io,caseyrygt/osf.io,Ghalko/osf.io,hmoco/osf.io,acshi/osf.io,brandonPurvis/osf.io,MerlinZhang/osf.io,SSJohns/osf.io,AndrewSallans/osf.io,leb2dg/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,cslzchen/osf.io,wearpants/osf.io,DanielSBrown/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,TomBaxter/osf.io,samanehsan/osf.io,kushG/osf.io,chrisseto/osf.io,binoculars/osf.io,arpitar/osf.io,doublebits/osf.io,ZobairAlijan/osf.io,samanehsan/osf.io,dplorimer/osf,acshi/osf.io,sbt9uc/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,felliott/osf.io,amyshi188/osf.io,mluke93/osf.io,zachjanicki/osf.io,GageGaskins/osf.io,dplorimer/osf,zkraime/osf.io,jolene-esposito/osf.io,rdhyee/osf.io,amyshi188/osf.io,doublebits/osf.io,lamdnhan/osf.io,mluo613/osf.io,Nesiehr/osf.io,brandonPurvis/osf.io,arpitar/osf.io,amyshi188/osf.io,jolene-esposito/osf.io,monikagrabowska/osf.io,zachjanicki/osf.io,petermalcolm/osf.io,cslzchen/osf.io,wearpants/osf.io,cldershem/osf.io,billyhunt/osf.io,njantrania/osf.io,DanielSBrown/osf.io,adlius/osf.io,jnayak1/osf.io,samanehsan/osf.io,felliott/osf.io,doublebits/osf.io,jolene-esposito/osf.io,crcresearch/osf.io,bdyetton/prettychart,bdyetton/prettychart,barbour-em/osf.io,lamdnhan/osf.io,Ghalko/osf.io,sbt9uc/osf.io,KAsante95/osf.io,sloria/osf.io,kch8qx/osf.io,erinspace/osf.io,zamattiac/osf.io,caneruguz/osf.io,kwierman/osf.io,samchrisinger/osf.io,TomHeatwole/osf.io,emetsger/osf.io,alexschiller/osf.io,doublebits/osf.io,caseyrygt/osf.io,ticklemepierce/osf.io,Nesiehr/osf.io,jinluyuan/osf.io,cosenal/osf.io,haoyuchen1992/osf.io,jmcarp/osf.io,kch8qx/osf.io,HarryRybacki/osf.io,MerlinZhang/osf.io,icereval/osf.io,kwierman/osf.io,alexschiller/osf.io,adlius/osf.io,icereval/osf.io,leb2dg/osf.io,jmcarp/osf.io,kch8qx/osf.io,jnayak1/osf.io,hmoco/osf.io,fabianvf/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,kch8qx/osf.io,CenterForOpenScience/osf.io,HarryRybacki/osf.io,bdyetton/prettychart,dplorimer/osf,aaxelb/osf.io,njantrania/osf.io,billyhunt/osf.io,ZobairAlijan/osf.io,rdhyee/osf.io,SSJohns/osf.io,billyhunt/osf.io,chennan47/osf.io,GaryKriebel/osf.io,chrisseto/osf.io,crcresearch/osf.io,revanthkolli/osf.io,kch8qx/osf.io,jnayak1/osf.io,danielneis/osf.io,GageGaskins/osf.io,amyshi188/osf.io,RomanZWang/osf.io,abought/osf.io,lyndsysimon/osf.io,cosenal/osf.io,KAsante95/osf.io,hmoco/osf.io,brandonPurvis/osf.io,haoyuchen1992/osf.io,mluke93/osf.io,GaryKriebel/osf.io,GaryKriebel/osf.io,reinaH/osf.io,leb2dg/osf.io,caneruguz/osf.io,emetsger/osf.io,jmcarp/osf.io,leb2dg/osf.io,mfraezz/osf.io,haoyuchen1992/osf.io,brandonPurvis/osf.io,cwisecarver/osf.io,caseyrollins/osf.io,caseyrygt/osf.io,kwierman/osf.io,AndrewSallans/osf.io,haoyuchen1992/osf.io,KAsante95/osf.io,jnayak1/osf.io,arpitar/osf.io,zamattiac/osf.io,petermalcolm/osf.io,HarryRybacki/osf.io,emetsger/osf.io,RomanZWang/osf.io,sloria/osf.io,erinspace/osf.io,fabianvf/osf.io,sloria/osf.io,acshi/osf.io,caseyrollins/osf.io,DanielSBrown/osf.io,asanfilippo7/osf.io,MerlinZhang/osf.io,himanshuo/osf.io,saradbowman/osf.io,alexschiller/osf.io,brianjgeiger/osf.io,sbt9uc/osf.io,rdhyee/osf.io,abought/osf.io,HarryRybacki/osf.io,TomHeatwole/osf.io,RomanZWang/osf.io,ZobairAlijan/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,abought/osf.io,lyndsysimon/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,ticklemepierce/osf.io,revanthkolli/osf.io,KAsante95/osf.io | from framework import StoredObject, fields
class Guid(StoredObject):
_id = fields.StringField()
referent = fields.AbstractForeignField()
_meta = {
'optimistic': True,
}
class GuidStoredObject(StoredObject):
# Redirect to content using URL redirect by default
redirect_mode = 'redirect'
def _ensure_guid(self):
"""Create GUID record if current record doesn't already have one, then
point GUID to self.
"""
# Create GUID with specified ID if ID provided
if self._primary_key:
# Done if GUID already exists
guid = Guid.load(self._primary_key)
if guid is not None:
return
# Create GUID
guid = Guid(
_id=self._primary_key,
referent=self
)
guid.save()
# Else create GUID optimistically
else:
# Create GUID
guid = Guid()
guid.save()
guid.referent = (guid._primary_key, self._name)
guid.save()
# Set primary key to GUID key
self._primary_key = guid._primary_key
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
- rv = super(GuidStoredObject, self).save(*args, **kwargs)
self._ensure_guid()
- return rv
+ return super(GuidStoredObject, self).save(*args, **kwargs)
@property
def annotations(self):
""" Get meta-data annotations associated with object. """
return self.metadata__annotated
| Fix last commit: Must ensure GUID before saving so that PK is defined | ## Code Before:
from framework import StoredObject, fields
class Guid(StoredObject):
_id = fields.StringField()
referent = fields.AbstractForeignField()
_meta = {
'optimistic': True,
}
class GuidStoredObject(StoredObject):
# Redirect to content using URL redirect by default
redirect_mode = 'redirect'
def _ensure_guid(self):
"""Create GUID record if current record doesn't already have one, then
point GUID to self.
"""
# Create GUID with specified ID if ID provided
if self._primary_key:
# Done if GUID already exists
guid = Guid.load(self._primary_key)
if guid is not None:
return
# Create GUID
guid = Guid(
_id=self._primary_key,
referent=self
)
guid.save()
# Else create GUID optimistically
else:
# Create GUID
guid = Guid()
guid.save()
guid.referent = (guid._primary_key, self._name)
guid.save()
# Set primary key to GUID key
self._primary_key = guid._primary_key
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
rv = super(GuidStoredObject, self).save(*args, **kwargs)
self._ensure_guid()
return rv
@property
def annotations(self):
""" Get meta-data annotations associated with object. """
return self.metadata__annotated
## Instruction:
Fix last commit: Must ensure GUID before saving so that PK is defined
## Code After:
from framework import StoredObject, fields
class Guid(StoredObject):
_id = fields.StringField()
referent = fields.AbstractForeignField()
_meta = {
'optimistic': True,
}
class GuidStoredObject(StoredObject):
# Redirect to content using URL redirect by default
redirect_mode = 'redirect'
def _ensure_guid(self):
"""Create GUID record if current record doesn't already have one, then
point GUID to self.
"""
# Create GUID with specified ID if ID provided
if self._primary_key:
# Done if GUID already exists
guid = Guid.load(self._primary_key)
if guid is not None:
return
# Create GUID
guid = Guid(
_id=self._primary_key,
referent=self
)
guid.save()
# Else create GUID optimistically
else:
# Create GUID
guid = Guid()
guid.save()
guid.referent = (guid._primary_key, self._name)
guid.save()
# Set primary key to GUID key
self._primary_key = guid._primary_key
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
self._ensure_guid()
return super(GuidStoredObject, self).save(*args, **kwargs)
@property
def annotations(self):
""" Get meta-data annotations associated with object. """
return self.metadata__annotated
| ...
def save(self, *args, **kwargs):
""" Ensure GUID on save initialization. """
self._ensure_guid()
return super(GuidStoredObject, self).save(*args, **kwargs)
@property
... |
73eacdde5067e60f40af000237d198748c5b3cc7 | PYNWapp/PYNWsite/models.py | PYNWapp/PYNWsite/models.py | from __future__ import unicode_literals
from django.db import models
from django.utils import timezone
# Create your models here.
class Event(models.Model):
name = models.CharField(max_length=200)
location = models.CharField(max_length=300)
event_date = models.DateTimeField('event date')
description = models.TextField()
def __str__(self):
return self.name
def is_future(self):
return self.event_date > timezone.now()
class Post(models.Model):
title = models.CharField(max_length=100, unique=True)
slug = models.SlugField(max_length=100, unique=True)
body = models.TextField()
posted = models.DateField(db_index=True, auto_now_add=True)
category = models.ForeignKey('Category')
class Category(models.Model):
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
| from __future__ import unicode_literals
from django.db import models
from django.utils import timezone
# Create your models here.
class Event(models.Model):
name = models.CharField(max_length=200)
location = models.CharField(max_length=300)
event_date = models.DateTimeField('event date')
description = models.TextField()
def __str__(self):
return self.name
def is_future(self):
return self.event_date > timezone.now()
class Post(models.Model):
title = models.CharField(max_length=100, unique=True)
slug = models.SlugField(max_length=100, unique=True)
body = models.TextField()
posted = models.DateField(db_index=True, auto_now_add=True)
category = models.ForeignKey('Category')
class Category(models.Model):
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
class Meta:
verbose_name_plural = 'Categories'
| Fix plural name for Categories model. | Fix plural name for Categories model.
| Python | mit | PythonNorthwestEngland/pynw-website,PythonNorthwestEngland/pynw-website | from __future__ import unicode_literals
from django.db import models
from django.utils import timezone
# Create your models here.
class Event(models.Model):
name = models.CharField(max_length=200)
location = models.CharField(max_length=300)
event_date = models.DateTimeField('event date')
description = models.TextField()
def __str__(self):
return self.name
def is_future(self):
return self.event_date > timezone.now()
class Post(models.Model):
title = models.CharField(max_length=100, unique=True)
slug = models.SlugField(max_length=100, unique=True)
body = models.TextField()
posted = models.DateField(db_index=True, auto_now_add=True)
category = models.ForeignKey('Category')
class Category(models.Model):
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
+ class Meta:
+ verbose_name_plural = 'Categories'
+ | Fix plural name for Categories model. | ## Code Before:
from __future__ import unicode_literals
from django.db import models
from django.utils import timezone
# Create your models here.
class Event(models.Model):
name = models.CharField(max_length=200)
location = models.CharField(max_length=300)
event_date = models.DateTimeField('event date')
description = models.TextField()
def __str__(self):
return self.name
def is_future(self):
return self.event_date > timezone.now()
class Post(models.Model):
title = models.CharField(max_length=100, unique=True)
slug = models.SlugField(max_length=100, unique=True)
body = models.TextField()
posted = models.DateField(db_index=True, auto_now_add=True)
category = models.ForeignKey('Category')
class Category(models.Model):
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
## Instruction:
Fix plural name for Categories model.
## Code After:
from __future__ import unicode_literals
from django.db import models
from django.utils import timezone
# Create your models here.
class Event(models.Model):
name = models.CharField(max_length=200)
location = models.CharField(max_length=300)
event_date = models.DateTimeField('event date')
description = models.TextField()
def __str__(self):
return self.name
def is_future(self):
return self.event_date > timezone.now()
class Post(models.Model):
title = models.CharField(max_length=100, unique=True)
slug = models.SlugField(max_length=100, unique=True)
body = models.TextField()
posted = models.DateField(db_index=True, auto_now_add=True)
category = models.ForeignKey('Category')
class Category(models.Model):
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
class Meta:
verbose_name_plural = 'Categories'
| ...
title = models.CharField(max_length=100, db_index=True)
slug = models.SlugField(max_length=100, db_index=True)
class Meta:
verbose_name_plural = 'Categories'
... |
109b753c807dae30ee736a6f071a058fa8b68d92 | tests/scoring_engine/web/views/test_services.py | tests/scoring_engine/web/views/test_services.py | from tests.scoring_engine.web.web_test import WebTest
class TestServices(WebTest):
def test_auth_required_services(self):
self.verify_auth_required('/services')
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
| from tests.scoring_engine.web.web_test import WebTest
from tests.scoring_engine.helpers import generate_sample_model_tree
class TestServices(WebTest):
def set_team_color(self, team, color):
team.color = color
self.session.add(team)
self.session.commit()
def set_blue_team(self, team):
self.set_team_color(team, 'Blue')
def set_white_team(self, team):
self.set_team_color(team, 'White')
def test_auth_required_services(self):
self.verify_auth_required('/services')
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
def test_normal_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 200
def test_unauthorized_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_white_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 302
def test_normal_service_id(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 200
def test_unauthorized_service_id(self):
self.create_default_user()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 302 | Update tests for services view | Update tests for services view
| Python | mit | pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine | from tests.scoring_engine.web.web_test import WebTest
+ from tests.scoring_engine.helpers import generate_sample_model_tree
class TestServices(WebTest):
+
+ def set_team_color(self, team, color):
+ team.color = color
+ self.session.add(team)
+ self.session.commit()
+
+ def set_blue_team(self, team):
+ self.set_team_color(team, 'Blue')
+
+ def set_white_team(self, team):
+ self.set_team_color(team, 'White')
def test_auth_required_services(self):
self.verify_auth_required('/services')
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
+ def test_normal_services(self):
+ user = self.create_default_user()
+ service = generate_sample_model_tree('Service', self.session)
+ self.set_blue_team(user.team)
+ service.team = user.team
+ self.session.add(service)
+ self.session.commit()
+ resp = self.auth_and_get_path('/services')
+ assert resp.status_code == 200
+
+ def test_unauthorized_services(self):
+ user = self.create_default_user()
+ service = generate_sample_model_tree('Service', self.session)
+ self.set_white_team(user.team)
+ service.team = user.team
+ self.session.add(service)
+ self.session.commit()
+ resp = self.auth_and_get_path('/services')
+ assert resp.status_code == 302
+
+ def test_normal_service_id(self):
+ user = self.create_default_user()
+ service = generate_sample_model_tree('Service', self.session)
+ self.set_blue_team(user.team)
+ service.team = user.team
+ self.session.add(service)
+ self.session.commit()
+ resp = self.auth_and_get_path('/service/1')
+ assert resp.status_code == 200
+
+ def test_unauthorized_service_id(self):
+ self.create_default_user()
+ resp = self.auth_and_get_path('/service/1')
+ assert resp.status_code == 302 | Update tests for services view | ## Code Before:
from tests.scoring_engine.web.web_test import WebTest
class TestServices(WebTest):
def test_auth_required_services(self):
self.verify_auth_required('/services')
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
## Instruction:
Update tests for services view
## Code After:
from tests.scoring_engine.web.web_test import WebTest
from tests.scoring_engine.helpers import generate_sample_model_tree
class TestServices(WebTest):
def set_team_color(self, team, color):
team.color = color
self.session.add(team)
self.session.commit()
def set_blue_team(self, team):
self.set_team_color(team, 'Blue')
def set_white_team(self, team):
self.set_team_color(team, 'White')
def test_auth_required_services(self):
self.verify_auth_required('/services')
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
def test_normal_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 200
def test_unauthorized_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_white_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 302
def test_normal_service_id(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 200
def test_unauthorized_service_id(self):
self.create_default_user()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 302 | ...
from tests.scoring_engine.web.web_test import WebTest
from tests.scoring_engine.helpers import generate_sample_model_tree
class TestServices(WebTest):
def set_team_color(self, team, color):
team.color = color
self.session.add(team)
self.session.commit()
def set_blue_team(self, team):
self.set_team_color(team, 'Blue')
def set_white_team(self, team):
self.set_team_color(team, 'White')
def test_auth_required_services(self):
...
def test_auth_required_service_id(self):
self.verify_auth_required('/service/1')
def test_normal_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 200
def test_unauthorized_services(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_white_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/services')
assert resp.status_code == 302
def test_normal_service_id(self):
user = self.create_default_user()
service = generate_sample_model_tree('Service', self.session)
self.set_blue_team(user.team)
service.team = user.team
self.session.add(service)
self.session.commit()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 200
def test_unauthorized_service_id(self):
self.create_default_user()
resp = self.auth_and_get_path('/service/1')
assert resp.status_code == 302
... |
f31ab02d9a409e31acf339db2b950216472b8e9e | salesforce/backend/operations.py | salesforce/backend/operations.py |
import re
from django.db.backends import BaseDatabaseOperations
"""
Default database operations, with unquoted names.
"""
class DatabaseOperations(BaseDatabaseOperations):
compiler_module = "salesforce.backend.compiler"
def __init__(self, connection):
# not calling superclass constructor to maintain Django 1.3 support
self.connection = connection
self._cache = None
def connection_init(self):
pass
def sql_flush(self, style, tables, sequences):
return []
def quote_name(self, name):
return name
def value_to_db_datetime(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def last_insert_id(self, cursor, db_table, db_column):
return cursor.lastrowid |
import re
from django.db.backends import BaseDatabaseOperations
"""
Default database operations, with unquoted names.
"""
class DatabaseOperations(BaseDatabaseOperations):
compiler_module = "salesforce.backend.compiler"
def __init__(self, connection):
# not calling superclass constructor to maintain Django 1.3 support
self.connection = connection
self._cache = None
def connection_init(self):
pass
def sql_flush(self, style, tables, sequences):
return []
def quote_name(self, name):
return name
def value_to_db_datetime(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def value_to_db_date(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def last_insert_id(self, cursor, db_table, db_column):
return cursor.lastrowid
| Fix bug with Date fields and SOQL. | Fix bug with Date fields and SOQL.
Fixes https://github.com/freelancersunion/django-salesforce/issues/10 | Python | mit | django-salesforce/django-salesforce,chromakey/django-salesforce,philchristensen/django-salesforce,hynekcer/django-salesforce,chromakey/django-salesforce,hynekcer/django-salesforce,hynekcer/django-salesforce,chromakey/django-salesforce,django-salesforce/django-salesforce,philchristensen/django-salesforce,django-salesforce/django-salesforce,philchristensen/django-salesforce |
import re
from django.db.backends import BaseDatabaseOperations
"""
Default database operations, with unquoted names.
"""
class DatabaseOperations(BaseDatabaseOperations):
compiler_module = "salesforce.backend.compiler"
def __init__(self, connection):
# not calling superclass constructor to maintain Django 1.3 support
self.connection = connection
self._cache = None
def connection_init(self):
pass
def sql_flush(self, style, tables, sequences):
return []
def quote_name(self, name):
return name
def value_to_db_datetime(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
+
+ def value_to_db_date(self, value):
+ """
+ We let the JSON serializer handle dates for us.
+ """
+ return value
def last_insert_id(self, cursor, db_table, db_column):
return cursor.lastrowid
+ | Fix bug with Date fields and SOQL. | ## Code Before:
import re
from django.db.backends import BaseDatabaseOperations
"""
Default database operations, with unquoted names.
"""
class DatabaseOperations(BaseDatabaseOperations):
compiler_module = "salesforce.backend.compiler"
def __init__(self, connection):
# not calling superclass constructor to maintain Django 1.3 support
self.connection = connection
self._cache = None
def connection_init(self):
pass
def sql_flush(self, style, tables, sequences):
return []
def quote_name(self, name):
return name
def value_to_db_datetime(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def last_insert_id(self, cursor, db_table, db_column):
return cursor.lastrowid
## Instruction:
Fix bug with Date fields and SOQL.
## Code After:
import re
from django.db.backends import BaseDatabaseOperations
"""
Default database operations, with unquoted names.
"""
class DatabaseOperations(BaseDatabaseOperations):
compiler_module = "salesforce.backend.compiler"
def __init__(self, connection):
# not calling superclass constructor to maintain Django 1.3 support
self.connection = connection
self._cache = None
def connection_init(self):
pass
def sql_flush(self, style, tables, sequences):
return []
def quote_name(self, name):
return name
def value_to_db_datetime(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def value_to_db_date(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def last_insert_id(self, cursor, db_table, db_column):
return cursor.lastrowid
| # ... existing code ...
"""
return value
def value_to_db_date(self, value):
"""
We let the JSON serializer handle dates for us.
"""
return value
def last_insert_id(self, cursor, db_table, db_column):
# ... rest of the code ... |
10e307a0dda94a9b38a1b7e143ef141e6062566b | skan/pipe.py | skan/pipe.py | from . import pre, csr
import imageio
import tqdm
import numpy as np
from skimage import morphology
import pandas as pd
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
image_format = (None if self.image_format.get() == 'auto'
else self.image_format.get())
results = []
from skan import pre, csr
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
if self.scale_metadata_path is not None:
md_path = self.scale_metadata_path.get().split(sep=',')
meta = image.meta
for key in md_path:
meta = meta[key]
scale = float(meta)
else:
scale = 1 # measurements will be in pixel units
pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() /
scale))
pixel_smoothing_radius = (self.smooth_radius.get() *
pixel_threshold_radius)
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
offset=self.brightness_offset.get())
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
framedata['squiggle'] = np.log2(framedata['branch-distance'] /
framedata['euclidean-distance'])
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
results = pd.concat(results)
| from . import pre, csr
import imageio
import tqdm
import numpy as np
from skimage import morphology
import pandas as pd
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
image_format = None if image_format == 'auto' else image_format
results = []
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
if scale_metadata_path is not None:
md_path = scale_metadata_path.split(sep=',')
meta = image.meta
for key in md_path:
meta = meta[key]
scale = float(meta)
else:
scale = 1 # measurements will be in pixel units
pixel_threshold_radius = int(np.ceil(threshold_radius / scale))
pixel_smoothing_radius = smooth_radius * pixel_threshold_radius
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
offset=brightness_offset)
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
framedata['squiggle'] = np.log2(framedata['branch-distance'] /
framedata['euclidean-distance'])
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
return pd.concat(results)
| Add module for start-to-finish functions | Add module for start-to-finish functions
| Python | bsd-3-clause | jni/skan | from . import pre, csr
import imageio
import tqdm
import numpy as np
from skimage import morphology
import pandas as pd
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
- image_format = (None if self.image_format.get() == 'auto'
+ image_format = None if image_format == 'auto' else image_format
- else self.image_format.get())
results = []
- from skan import pre, csr
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
- if self.scale_metadata_path is not None:
+ if scale_metadata_path is not None:
- md_path = self.scale_metadata_path.get().split(sep=',')
+ md_path = scale_metadata_path.split(sep=',')
meta = image.meta
for key in md_path:
meta = meta[key]
scale = float(meta)
else:
scale = 1 # measurements will be in pixel units
- pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() /
+ pixel_threshold_radius = int(np.ceil(threshold_radius / scale))
+ pixel_smoothing_radius = smooth_radius * pixel_threshold_radius
- scale))
- pixel_smoothing_radius = (self.smooth_radius.get() *
- pixel_threshold_radius)
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
- offset=self.brightness_offset.get())
+ offset=brightness_offset)
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
framedata['squiggle'] = np.log2(framedata['branch-distance'] /
framedata['euclidean-distance'])
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
- results = pd.concat(results)
+ return pd.concat(results)
| Add module for start-to-finish functions | ## Code Before:
from . import pre, csr
import imageio
import tqdm
import numpy as np
from skimage import morphology
import pandas as pd
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
image_format = (None if self.image_format.get() == 'auto'
else self.image_format.get())
results = []
from skan import pre, csr
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
if self.scale_metadata_path is not None:
md_path = self.scale_metadata_path.get().split(sep=',')
meta = image.meta
for key in md_path:
meta = meta[key]
scale = float(meta)
else:
scale = 1 # measurements will be in pixel units
pixel_threshold_radius = int(np.ceil(self.threshold_radius.get() /
scale))
pixel_smoothing_radius = (self.smooth_radius.get() *
pixel_threshold_radius)
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
offset=self.brightness_offset.get())
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
framedata['squiggle'] = np.log2(framedata['branch-distance'] /
framedata['euclidean-distance'])
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
results = pd.concat(results)
## Instruction:
Add module for start-to-finish functions
## Code After:
from . import pre, csr
import imageio
import tqdm
import numpy as np
from skimage import morphology
import pandas as pd
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
image_format = None if image_format == 'auto' else image_format
results = []
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
if scale_metadata_path is not None:
md_path = scale_metadata_path.split(sep=',')
meta = image.meta
for key in md_path:
meta = meta[key]
scale = float(meta)
else:
scale = 1 # measurements will be in pixel units
pixel_threshold_radius = int(np.ceil(threshold_radius / scale))
pixel_smoothing_radius = smooth_radius * pixel_threshold_radius
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
offset=brightness_offset)
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
framedata['squiggle'] = np.log2(framedata['branch-distance'] /
framedata['euclidean-distance'])
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
return pd.concat(results)
| // ... existing code ...
def process_images(filenames, image_format, threshold_radius,
smooth_radius, brightness_offset, scale_metadata_path):
image_format = None if image_format == 'auto' else image_format
results = []
for file in tqdm(filenames):
image = imageio.imread(file, format=image_format)
if scale_metadata_path is not None:
md_path = scale_metadata_path.split(sep=',')
meta = image.meta
for key in md_path:
// ... modified code ...
else:
scale = 1 # measurements will be in pixel units
pixel_threshold_radius = int(np.ceil(threshold_radius / scale))
pixel_smoothing_radius = smooth_radius * pixel_threshold_radius
thresholded = pre.threshold(image, sigma=pixel_smoothing_radius,
radius=pixel_threshold_radius,
offset=brightness_offset)
skeleton = morphology.skeletonize(thresholded)
framedata = csr.summarise(skeleton, spacing=scale)
...
framedata['filename'] = [file] * len(framedata)
results.append(framedata)
return pd.concat(results)
// ... rest of the code ... |
b56712563e4205ccbf8b98deace4197e2f250361 | movement.py | movement.py | if __name__ == "__main__":
x, y = 0, 0
steps = 0
while True:
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
if dir in directions:
print("You moved %s, " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
break
else:
print("You tried to move %s but that is not possible." % dir) | if __name__ == "__main__":
x, y = 0, 0
steps = 0
while True:
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
abb_directions = {'n': (0, 1),
's' : (0, -1),
'e' : (1, 0),
'w' : (-1, 0)}
long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'}
dir = dir.lower().replace(" ", "")
if dir in directions:
print("You moved %s. " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
elif dir in abb_directions:
print("You moved %s. " % long_directions[dir])
x += abb_directions[dir][0]
y += abb_directions[dir][1]
steps += 1
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
break
else:
print("You tried to move %s but that is not possible." % dir) | Add abbreviations and space handling | Add abbreviations and space handling
| Python | mit | mewturn/Python | if __name__ == "__main__":
x, y = 0, 0
steps = 0
while True:
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
- directions = { 'north': (0, 1),
+ directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
-
+
+ abb_directions = {'n': (0, 1),
+ 's' : (0, -1),
+ 'e' : (1, 0),
+ 'w' : (-1, 0)}
+
+ long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'}
+
+ dir = dir.lower().replace(" ", "")
if dir in directions:
- print("You moved %s, " % dir)
+ print("You moved %s. " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
-
+
+ elif dir in abb_directions:
+ print("You moved %s. " % long_directions[dir])
+ x += abb_directions[dir][0]
+ y += abb_directions[dir][1]
+ steps += 1
+
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
break
else:
print("You tried to move %s but that is not possible." % dir) | Add abbreviations and space handling | ## Code Before:
if __name__ == "__main__":
x, y = 0, 0
steps = 0
while True:
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
if dir in directions:
print("You moved %s, " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
break
else:
print("You tried to move %s but that is not possible." % dir)
## Instruction:
Add abbreviations and space handling
## Code After:
if __name__ == "__main__":
x, y = 0, 0
steps = 0
while True:
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
abb_directions = {'n': (0, 1),
's' : (0, -1),
'e' : (1, 0),
'w' : (-1, 0)}
long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'}
dir = dir.lower().replace(" ", "")
if dir in directions:
print("You moved %s. " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
elif dir in abb_directions:
print("You moved %s. " % long_directions[dir])
x += abb_directions[dir][0]
y += abb_directions[dir][1]
steps += 1
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
break
else:
print("You tried to move %s but that is not possible." % dir) | // ... existing code ...
dir = input('Your current position is %s, %s, where would you like to move to? ' % (str(x), str(y)))
directions = { 'north': (0, 1),
'south' : (0, -1),
'east' : (1, 0),
'west' : (-1, 0)}
abb_directions = {'n': (0, 1),
's' : (0, -1),
'e' : (1, 0),
'w' : (-1, 0)}
long_directions = {'n' : 'north', 's' : 'south', 'e' : 'east', 'w' : 'west'}
dir = dir.lower().replace(" ", "")
if dir in directions:
print("You moved %s. " % dir)
x += directions[dir][0]
y += directions[dir][1]
steps += 1
elif dir in abb_directions:
print("You moved %s. " % long_directions[dir])
x += abb_directions[dir][0]
y += abb_directions[dir][1]
steps += 1
elif dir == "leave":
print("Goodbye. You moved %s steps in total." % steps)
// ... rest of the code ... |
a0aa74d9e6295e34f02b4eefd76e7eb9a1e6425f | node/floor_divide.py | node/floor_divide.py |
from nodes import Node
class FloorDiv(Node):
char = "f"
args = 2
results = 1
@Node.test_func([3,2], [1])
@Node.test_func([6,-3], [-2])
def func(self, a:Node.number,b:Node.number):
"""a/b. Rounds down, returns an int."""
return a//b
@Node.test_func(["test", "e"], [["t", "e", "st"]])
def partition(self, string:str, sep:str):
"""Split the string at the first occurrence of sep,
return a 3-list containing the part before the separator,
the separator itself, and the part after the separator.
If the separator is not found,
return a 3-list containing the string itself,
followed by two empty strings."""
return [list(string.partition(sep))] |
from nodes import Node
class FloorDiv(Node):
char = "f"
args = 2
results = 1
@Node.test_func([3,2], [1])
@Node.test_func([6,-3], [-2])
def func(self, a:Node.number,b:Node.number):
"""a/b. Rounds down, returns an int."""
return a//b
@Node.test_func(["test", "e"], [["t", "e", "st"]])
def partition(self, string:str, sep:str):
"""Split the string at the first occurrence of sep,
return a 3-list containing the part before the separator,
the separator itself, and the part after the separator.
If the separator is not found,
return a 3-list containing the string itself,
followed by two empty strings."""
return [list(string.partition(sep))]
@Node.test_func(["134", 1], [["134"]])
@Node.test_func(["1234", 2], [["12", "34"]])
@Node.test_func(["1234", 3], [["1", "2", "34"]])
@Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]])
def chunk(self, inp:Node.indexable, num:Node.number):
"""Return inp seperated into num groups"""
rtn = []
last = 0
size = len(inp)//num
for i in range(size, len(inp), size):
rtn.append(inp[last:i])
last = i
if len(rtn) != num:
rtn.append(inp[last:])
else:
rtn[-1] += inp[last:]
if len(rtn):
if isinstance(inp, str):
rtn[-1] = "".join(rtn[-1])
else:
rtn[-1] = type(inp)(rtn[-1])
return [rtn] | Add a group chunk, chunks a list into N groups | Add a group chunk, chunks a list into N groups
| Python | mit | muddyfish/PYKE,muddyfish/PYKE |
from nodes import Node
class FloorDiv(Node):
char = "f"
args = 2
results = 1
@Node.test_func([3,2], [1])
@Node.test_func([6,-3], [-2])
def func(self, a:Node.number,b:Node.number):
"""a/b. Rounds down, returns an int."""
return a//b
@Node.test_func(["test", "e"], [["t", "e", "st"]])
def partition(self, string:str, sep:str):
"""Split the string at the first occurrence of sep,
return a 3-list containing the part before the separator,
the separator itself, and the part after the separator.
If the separator is not found,
return a 3-list containing the string itself,
followed by two empty strings."""
return [list(string.partition(sep))]
+
+ @Node.test_func(["134", 1], [["134"]])
+ @Node.test_func(["1234", 2], [["12", "34"]])
+ @Node.test_func(["1234", 3], [["1", "2", "34"]])
+ @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]])
+ def chunk(self, inp:Node.indexable, num:Node.number):
+ """Return inp seperated into num groups"""
+ rtn = []
+ last = 0
+ size = len(inp)//num
+ for i in range(size, len(inp), size):
+ rtn.append(inp[last:i])
+ last = i
+ if len(rtn) != num:
+ rtn.append(inp[last:])
+ else:
+ rtn[-1] += inp[last:]
+ if len(rtn):
+ if isinstance(inp, str):
+ rtn[-1] = "".join(rtn[-1])
+ else:
+ rtn[-1] = type(inp)(rtn[-1])
+ return [rtn] | Add a group chunk, chunks a list into N groups | ## Code Before:
from nodes import Node
class FloorDiv(Node):
char = "f"
args = 2
results = 1
@Node.test_func([3,2], [1])
@Node.test_func([6,-3], [-2])
def func(self, a:Node.number,b:Node.number):
"""a/b. Rounds down, returns an int."""
return a//b
@Node.test_func(["test", "e"], [["t", "e", "st"]])
def partition(self, string:str, sep:str):
"""Split the string at the first occurrence of sep,
return a 3-list containing the part before the separator,
the separator itself, and the part after the separator.
If the separator is not found,
return a 3-list containing the string itself,
followed by two empty strings."""
return [list(string.partition(sep))]
## Instruction:
Add a group chunk, chunks a list into N groups
## Code After:
from nodes import Node
class FloorDiv(Node):
char = "f"
args = 2
results = 1
@Node.test_func([3,2], [1])
@Node.test_func([6,-3], [-2])
def func(self, a:Node.number,b:Node.number):
"""a/b. Rounds down, returns an int."""
return a//b
@Node.test_func(["test", "e"], [["t", "e", "st"]])
def partition(self, string:str, sep:str):
"""Split the string at the first occurrence of sep,
return a 3-list containing the part before the separator,
the separator itself, and the part after the separator.
If the separator is not found,
return a 3-list containing the string itself,
followed by two empty strings."""
return [list(string.partition(sep))]
@Node.test_func(["134", 1], [["134"]])
@Node.test_func(["1234", 2], [["12", "34"]])
@Node.test_func(["1234", 3], [["1", "2", "34"]])
@Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]])
def chunk(self, inp:Node.indexable, num:Node.number):
"""Return inp seperated into num groups"""
rtn = []
last = 0
size = len(inp)//num
for i in range(size, len(inp), size):
rtn.append(inp[last:i])
last = i
if len(rtn) != num:
rtn.append(inp[last:])
else:
rtn[-1] += inp[last:]
if len(rtn):
if isinstance(inp, str):
rtn[-1] = "".join(rtn[-1])
else:
rtn[-1] = type(inp)(rtn[-1])
return [rtn] | // ... existing code ...
followed by two empty strings."""
return [list(string.partition(sep))]
@Node.test_func(["134", 1], [["134"]])
@Node.test_func(["1234", 2], [["12", "34"]])
@Node.test_func(["1234", 3], [["1", "2", "34"]])
@Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]])
def chunk(self, inp:Node.indexable, num:Node.number):
"""Return inp seperated into num groups"""
rtn = []
last = 0
size = len(inp)//num
for i in range(size, len(inp), size):
rtn.append(inp[last:i])
last = i
if len(rtn) != num:
rtn.append(inp[last:])
else:
rtn[-1] += inp[last:]
if len(rtn):
if isinstance(inp, str):
rtn[-1] = "".join(rtn[-1])
else:
rtn[-1] = type(inp)(rtn[-1])
return [rtn]
// ... rest of the code ... |
f353ee5d2e2cf5fd4ee86776fc7e5ee6cb8a3238 | sierra_adapter/build_windows.py | sierra_adapter/build_windows.py |
import datetime as dt
import json
import boto3
import docopt
import maya
args = docopt.docopt(__doc__)
start = maya.parse(args['--start']).datetime()
end = maya.parse(args['--end']).datetime()
minutes = int(args['--interval'] or 30)
resource = args['--resource']
assert resource in ('bibs', 'items')
def generate_windows(start, end, minutes):
current = start
while current <= end:
yield {
'start': current.isoformat(),
'end': (current + dt.timedelta(minutes=minutes)).isoformat(),
}
current += dt.timedelta(minutes=minutes - 1)
client = boto3.client('sns')
for window in generate_windows(start, end, minutes):
resp = client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
Subject=f'Window sent by {__file__}'
)
print(resp)
|
import datetime as dt
import json
import math
import boto3
import docopt
import maya
import tqdm
args = docopt.docopt(__doc__)
start = maya.parse(args['--start']).datetime()
end = maya.parse(args['--end']).datetime()
minutes = int(args['--interval'] or 30)
resource = args['--resource']
assert resource in ('bibs', 'items')
def generate_windows(start, end, minutes):
current = start
while current <= end:
yield {
'start': current.isoformat(),
'end': (current + dt.timedelta(minutes=minutes)).isoformat(),
}
current += dt.timedelta(minutes=minutes - 1)
client = boto3.client('sns')
for window in tqdm.tqdm(
generate_windows(start, end, minutes),
total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1))
):
client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
Subject=f'Window sent by {__file__}'
)
| Print a progress meter when pushing windows | Print a progress meter when pushing windows
| Python | mit | wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api |
import datetime as dt
import json
+ import math
import boto3
import docopt
import maya
+ import tqdm
args = docopt.docopt(__doc__)
start = maya.parse(args['--start']).datetime()
end = maya.parse(args['--end']).datetime()
minutes = int(args['--interval'] or 30)
resource = args['--resource']
assert resource in ('bibs', 'items')
def generate_windows(start, end, minutes):
current = start
while current <= end:
yield {
'start': current.isoformat(),
'end': (current + dt.timedelta(minutes=minutes)).isoformat(),
}
current += dt.timedelta(minutes=minutes - 1)
client = boto3.client('sns')
+ for window in tqdm.tqdm(
- for window in generate_windows(start, end, minutes):
+ generate_windows(start, end, minutes),
+ total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1))
+ ):
- resp = client.publish(
+ client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
Subject=f'Window sent by {__file__}'
)
- print(resp)
| Print a progress meter when pushing windows | ## Code Before:
import datetime as dt
import json
import boto3
import docopt
import maya
args = docopt.docopt(__doc__)
start = maya.parse(args['--start']).datetime()
end = maya.parse(args['--end']).datetime()
minutes = int(args['--interval'] or 30)
resource = args['--resource']
assert resource in ('bibs', 'items')
def generate_windows(start, end, minutes):
current = start
while current <= end:
yield {
'start': current.isoformat(),
'end': (current + dt.timedelta(minutes=minutes)).isoformat(),
}
current += dt.timedelta(minutes=minutes - 1)
client = boto3.client('sns')
for window in generate_windows(start, end, minutes):
resp = client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
Subject=f'Window sent by {__file__}'
)
print(resp)
## Instruction:
Print a progress meter when pushing windows
## Code After:
import datetime as dt
import json
import math
import boto3
import docopt
import maya
import tqdm
args = docopt.docopt(__doc__)
start = maya.parse(args['--start']).datetime()
end = maya.parse(args['--end']).datetime()
minutes = int(args['--interval'] or 30)
resource = args['--resource']
assert resource in ('bibs', 'items')
def generate_windows(start, end, minutes):
current = start
while current <= end:
yield {
'start': current.isoformat(),
'end': (current + dt.timedelta(minutes=minutes)).isoformat(),
}
current += dt.timedelta(minutes=minutes - 1)
client = boto3.client('sns')
for window in tqdm.tqdm(
generate_windows(start, end, minutes),
total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1))
):
client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
Subject=f'Window sent by {__file__}'
)
| // ... existing code ...
import datetime as dt
import json
import math
import boto3
// ... modified code ...
import docopt
import maya
import tqdm
args = docopt.docopt(__doc__)
...
client = boto3.client('sns')
for window in tqdm.tqdm(
generate_windows(start, end, minutes),
total=math.ceil((end - start).total_seconds() / 60 / (minutes - 1))
):
client.publish(
TopicArn=f'arn:aws:sns:eu-west-1:760097843905:sierra_{resource}_windows',
Message=json.dumps(window),
...
Subject=f'Window sent by {__file__}'
)
// ... rest of the code ... |
b419e78a42e7b8f073bc5d9502dffc97c5d627fb | apps/chats/forms.py | apps/chats/forms.py | from django import forms
from django.contrib.auth.models import User
from chats.models import Chat
from profiles.models import FriendGroup
class PublicChatForm(forms.ModelForm):
"""Public-facing Chat form used in the web-interface for users."""
class Meta:
fields = ('text',)
model = Chat
| from django import forms
from django.contrib.auth.models import User
from chats.models import Chat
from profiles.models import FriendGroup
class PublicChatForm(forms.ModelForm):
"""Public-facing Chat form used in the web-interface for users."""
class Meta:
fields = (
'friend_groups',
'text',
)
model = Chat
| Add friend_groups to the ChatForm | Add friend_groups to the ChatForm
| Python | mit | tofumatt/quotes,tofumatt/quotes | from django import forms
from django.contrib.auth.models import User
from chats.models import Chat
from profiles.models import FriendGroup
class PublicChatForm(forms.ModelForm):
"""Public-facing Chat form used in the web-interface for users."""
class Meta:
- fields = ('text',)
+ fields = (
+ 'friend_groups',
+ 'text',
+ )
model = Chat
| Add friend_groups to the ChatForm | ## Code Before:
from django import forms
from django.contrib.auth.models import User
from chats.models import Chat
from profiles.models import FriendGroup
class PublicChatForm(forms.ModelForm):
"""Public-facing Chat form used in the web-interface for users."""
class Meta:
fields = ('text',)
model = Chat
## Instruction:
Add friend_groups to the ChatForm
## Code After:
from django import forms
from django.contrib.auth.models import User
from chats.models import Chat
from profiles.models import FriendGroup
class PublicChatForm(forms.ModelForm):
"""Public-facing Chat form used in the web-interface for users."""
class Meta:
fields = (
'friend_groups',
'text',
)
model = Chat
| ...
class Meta:
fields = (
'friend_groups',
'text',
)
model = Chat
... |
27b9bd22bb43b8b86ae1c40a90c1fae7157dcb86 | app/tests.py | app/tests.py | from app.test_base import BaseTestCase
class TestTopLevelFunctions(BaseTestCase):
def test_index_response(self):
response = self.client.get('/')
self.assert200(response)
| from app.test_base import BaseTestCase
class TestTopLevelFunctions(BaseTestCase):
def test_index_response(self):
response = self.client.get('/')
self.assert200(response)
def test_login_required(self):
self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd')
self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew')
self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew')
self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew')
self.check_login_required('/settings', '/login?next=%2Fsettings')
self.check_login_required('/review', '/login?next=%2Freview')
self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew')
self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs')
def check_login_required(self, attempted_location, redirected_location):
response = self.client.get(attempted_location)
self.assertTrue(response.status_code in (301, 302))
self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
| Add test to verify login required for protected pages | Add test to verify login required for protected pages
| Python | mit | rtfoley/scorepy,rtfoley/scorepy,rtfoley/scorepy | from app.test_base import BaseTestCase
class TestTopLevelFunctions(BaseTestCase):
def test_index_response(self):
response = self.client.get('/')
self.assert200(response)
+ def test_login_required(self):
+ self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd')
+ self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew')
+ self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew')
+ self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew')
+ self.check_login_required('/settings', '/login?next=%2Fsettings')
+ self.check_login_required('/review', '/login?next=%2Freview')
+ self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew')
+ self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs')
+
+ def check_login_required(self, attempted_location, redirected_location):
+ response = self.client.get(attempted_location)
+ self.assertTrue(response.status_code in (301, 302))
+ self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
+ | Add test to verify login required for protected pages | ## Code Before:
from app.test_base import BaseTestCase
class TestTopLevelFunctions(BaseTestCase):
def test_index_response(self):
response = self.client.get('/')
self.assert200(response)
## Instruction:
Add test to verify login required for protected pages
## Code After:
from app.test_base import BaseTestCase
class TestTopLevelFunctions(BaseTestCase):
def test_index_response(self):
response = self.client.get('/')
self.assert200(response)
def test_login_required(self):
self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd')
self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew')
self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew')
self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew')
self.check_login_required('/settings', '/login?next=%2Fsettings')
self.check_login_required('/review', '/login?next=%2Freview')
self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew')
self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs')
def check_login_required(self, attempted_location, redirected_location):
response = self.client.get(attempted_location)
self.assertTrue(response.status_code in (301, 302))
self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
| // ... existing code ...
response = self.client.get('/')
self.assert200(response)
def test_login_required(self):
self.check_login_required('/scores/add', '/login?next=%2Fscores%2Fadd')
self.check_login_required('/judging/presentation/new', '/login?next=%2Fjudging%2Fpresentation%2Fnew')
self.check_login_required('/judging/technical/new', '/login?next=%2Fjudging%2Ftechnical%2Fnew')
self.check_login_required('/judging/core_values/new', '/login?next=%2Fjudging%2Fcore_values%2Fnew')
self.check_login_required('/settings', '/login?next=%2Fsettings')
self.check_login_required('/review', '/login?next=%2Freview')
self.check_login_required('/teams/new', '/login?next=%2Fteams%2Fnew')
self.check_login_required('/scores/playoffs', '/login?next=%2Fscores%2Fplayoffs')
def check_login_required(self, attempted_location, redirected_location):
response = self.client.get(attempted_location)
self.assertTrue(response.status_code in (301, 302))
self.assertEqual(response.location, 'http://' + self.app.config['SERVER_NAME'] + redirected_location)
// ... rest of the code ... |
c6298a573dc3188b8c57954287d78e7da253483a | lot/urls.py | lot/urls.py |
from django.conf.urls import patterns, url
from . import views
urlpatterns = patterns("",
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
)
|
from django.conf.urls import url
from . import views
urlpatterns = [
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
]
| Update to new-style urlpatterns format | Update to new-style urlpatterns format
| Python | bsd-3-clause | ABASystems/django-lot |
- from django.conf.urls import patterns, url
+ from django.conf.urls import url
from . import views
- urlpatterns = patterns("",
+ urlpatterns = [
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
- )
+ ]
| Update to new-style urlpatterns format | ## Code Before:
from django.conf.urls import patterns, url
from . import views
urlpatterns = patterns("",
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
)
## Instruction:
Update to new-style urlpatterns format
## Code After:
from django.conf.urls import url
from . import views
urlpatterns = [
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
]
| // ... existing code ...
from django.conf.urls import url
from . import views
urlpatterns = [
url(r"^login/(?P<uuid>[\da-f]{8}-([\da-f]{4}-){3}[\da-f]{12})/$", views.LOTLogin.as_view(), name="login"),
]
// ... rest of the code ... |
9b10f600b5611380f72fe2aeacfe2ee6f02e4e3a | kicad_footprint_load.py | kicad_footprint_load.py | import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
#Ignore paths with unicode as KiCad can't deal with them in enumerate
list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
| import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
list_of_footprints = src_plugin.FootprintEnumerate(libpath)
| Switch to old invocation of FootprintEnumerate | Switch to old invocation of FootprintEnumerate
| Python | mit | monostable/haskell-kicad-data,monostable/haskell-kicad-data,kasbah/haskell-kicad-data | import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
- #Ignore paths with unicode as KiCad can't deal with them in enumerate
- list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
+ list_of_footprints = src_plugin.FootprintEnumerate(libpath)
| Switch to old invocation of FootprintEnumerate | ## Code Before:
import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
#Ignore paths with unicode as KiCad can't deal with them in enumerate
list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
## Instruction:
Switch to old invocation of FootprintEnumerate
## Code After:
import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
list_of_footprints = src_plugin.FootprintEnumerate(libpath)
| # ... existing code ...
for libpath in pretties:
list_of_footprints = src_plugin.FootprintEnumerate(libpath)
# ... rest of the code ... |
32c40710a562b194385f2340bf882cb3709b74e3 | masquerade/urls.py | masquerade/urls.py | from django.conf.urls import patterns, url
urlpatterns = patterns('',
url(r'^mask/$', 'masquerade.views.mask'),
url(r'^unmask/$', 'masquerade.views.unmask'),
)
| from django.conf.urls import patterns, url
from masquerade.views import mask
from masquerade.views import unmask
urlpatterns = [
url(r'^mask/$', mask),
url(r'^unmask/$', unmask),
]
| Fix Django 1.10 deprecation warning | Fix Django 1.10 deprecation warning
| Python | apache-2.0 | erikcw/django-masquerade,erikcw/django-masquerade,erikcw/django-masquerade | from django.conf.urls import patterns, url
+ from masquerade.views import mask
+ from masquerade.views import unmask
- urlpatterns = patterns('',
- url(r'^mask/$', 'masquerade.views.mask'),
- url(r'^unmask/$', 'masquerade.views.unmask'),
- )
+
+ urlpatterns = [
+ url(r'^mask/$', mask),
+ url(r'^unmask/$', unmask),
+ ]
+ | Fix Django 1.10 deprecation warning | ## Code Before:
from django.conf.urls import patterns, url
urlpatterns = patterns('',
url(r'^mask/$', 'masquerade.views.mask'),
url(r'^unmask/$', 'masquerade.views.unmask'),
)
## Instruction:
Fix Django 1.10 deprecation warning
## Code After:
from django.conf.urls import patterns, url
from masquerade.views import mask
from masquerade.views import unmask
urlpatterns = [
url(r'^mask/$', mask),
url(r'^unmask/$', unmask),
]
| // ... existing code ...
from django.conf.urls import patterns, url
from masquerade.views import mask
from masquerade.views import unmask
urlpatterns = [
url(r'^mask/$', mask),
url(r'^unmask/$', unmask),
]
// ... rest of the code ... |
07455e5821d21c988c7c5fcda9345e99355eb4e7 | redash/__init__.py | redash/__init__.py | import json
import urlparse
from flask import Flask, make_response
from flask.ext.restful import Api
from flask_peewee.db import Database
import redis
from redash import settings, utils
__version__ = '0.3.2'
app = Flask(__name__,
template_folder=settings.STATIC_ASSETS_PATH,
static_folder=settings.STATIC_ASSETS_PATH,
static_path='/static')
api = Api(app)
# configure our database
settings.DATABASE_CONFIG.update({'threadlocals': True})
app.config['DATABASE'] = settings.DATABASE_CONFIG
db = Database(app)
from redash.authentication import setup_authentication
auth = setup_authentication(app)
@api.representation('application/json')
def json_representation(data, code, headers=None):
resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code)
resp.headers.extend(headers or {})
return resp
redis_url = urlparse.urlparse(settings.REDIS_URL)
redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password)
from redash import data
data_manager = data.Manager(redis_connection, db)
from redash import controllers | import json
import urlparse
from flask import Flask, make_response
from flask.ext.restful import Api
from flask_peewee.db import Database
import redis
from redash import settings, utils
__version__ = '0.3.2'
app = Flask(__name__,
template_folder=settings.STATIC_ASSETS_PATH,
static_folder=settings.STATIC_ASSETS_PATH,
static_path='/static')
api = Api(app)
# configure our database
settings.DATABASE_CONFIG.update({'threadlocals': True})
app.config['DATABASE'] = settings.DATABASE_CONFIG
db = Database(app)
from redash.authentication import setup_authentication
auth = setup_authentication(app)
@api.representation('application/json')
def json_representation(data, code, headers=None):
resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code)
resp.headers.extend(headers or {})
return resp
redis_url = urlparse.urlparse(settings.REDIS_URL)
if redis_url.path:
redis_db = redis_url.path[1]
else:
redis_db = 0
redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password)
from redash import data
data_manager = data.Manager(redis_connection, db)
from redash import controllers | Use database number from redis url if available. | Use database number from redis url if available.
| Python | bsd-2-clause | chriszs/redash,imsally/redash,44px/redash,guaguadev/redash,denisov-vlad/redash,rockwotj/redash,44px/redash,rockwotj/redash,getredash/redash,ninneko/redash,akariv/redash,amino-data/redash,akariv/redash,imsally/redash,EverlyWell/redash,getredash/redash,easytaxibr/redash,M32Media/redash,vishesh92/redash,ninneko/redash,getredash/redash,easytaxibr/redash,crowdworks/redash,hudl/redash,ninneko/redash,denisov-vlad/redash,stefanseifert/redash,amino-data/redash,useabode/redash,guaguadev/redash,jmvasquez/redashtest,alexanderlz/redash,moritz9/redash,pubnative/redash,denisov-vlad/redash,imsally/redash,rockwotj/redash,pubnative/redash,vishesh92/redash,useabode/redash,crowdworks/redash,akariv/redash,44px/redash,easytaxibr/redash,getredash/redash,chriszs/redash,akariv/redash,guaguadev/redash,44px/redash,stefanseifert/redash,denisov-vlad/redash,amino-data/redash,moritz9/redash,M32Media/redash,crowdworks/redash,M32Media/redash,M32Media/redash,alexanderlz/redash,crowdworks/redash,jmvasquez/redashtest,stefanseifert/redash,denisov-vlad/redash,pubnative/redash,EverlyWell/redash,moritz9/redash,stefanseifert/redash,stefanseifert/redash,pubnative/redash,imsally/redash,EverlyWell/redash,akariv/redash,ninneko/redash,moritz9/redash,jmvasquez/redashtest,useabode/redash,vishesh92/redash,chriszs/redash,getredash/redash,jmvasquez/redashtest,useabode/redash,easytaxibr/redash,ninneko/redash,amino-data/redash,pubnative/redash,hudl/redash,EverlyWell/redash,guaguadev/redash,jmvasquez/redashtest,easytaxibr/redash,hudl/redash,vishesh92/redash,chriszs/redash,rockwotj/redash,alexanderlz/redash,hudl/redash,alexanderlz/redash,guaguadev/redash | import json
import urlparse
from flask import Flask, make_response
from flask.ext.restful import Api
from flask_peewee.db import Database
import redis
from redash import settings, utils
__version__ = '0.3.2'
app = Flask(__name__,
template_folder=settings.STATIC_ASSETS_PATH,
static_folder=settings.STATIC_ASSETS_PATH,
static_path='/static')
api = Api(app)
# configure our database
settings.DATABASE_CONFIG.update({'threadlocals': True})
app.config['DATABASE'] = settings.DATABASE_CONFIG
db = Database(app)
from redash.authentication import setup_authentication
auth = setup_authentication(app)
@api.representation('application/json')
def json_representation(data, code, headers=None):
resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code)
resp.headers.extend(headers or {})
return resp
redis_url = urlparse.urlparse(settings.REDIS_URL)
+ if redis_url.path:
+ redis_db = redis_url.path[1]
+ else:
+ redis_db = 0
- redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password)
+ redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password)
from redash import data
data_manager = data.Manager(redis_connection, db)
from redash import controllers | Use database number from redis url if available. | ## Code Before:
import json
import urlparse
from flask import Flask, make_response
from flask.ext.restful import Api
from flask_peewee.db import Database
import redis
from redash import settings, utils
__version__ = '0.3.2'
app = Flask(__name__,
template_folder=settings.STATIC_ASSETS_PATH,
static_folder=settings.STATIC_ASSETS_PATH,
static_path='/static')
api = Api(app)
# configure our database
settings.DATABASE_CONFIG.update({'threadlocals': True})
app.config['DATABASE'] = settings.DATABASE_CONFIG
db = Database(app)
from redash.authentication import setup_authentication
auth = setup_authentication(app)
@api.representation('application/json')
def json_representation(data, code, headers=None):
resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code)
resp.headers.extend(headers or {})
return resp
redis_url = urlparse.urlparse(settings.REDIS_URL)
redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=0, password=redis_url.password)
from redash import data
data_manager = data.Manager(redis_connection, db)
from redash import controllers
## Instruction:
Use database number from redis url if available.
## Code After:
import json
import urlparse
from flask import Flask, make_response
from flask.ext.restful import Api
from flask_peewee.db import Database
import redis
from redash import settings, utils
__version__ = '0.3.2'
app = Flask(__name__,
template_folder=settings.STATIC_ASSETS_PATH,
static_folder=settings.STATIC_ASSETS_PATH,
static_path='/static')
api = Api(app)
# configure our database
settings.DATABASE_CONFIG.update({'threadlocals': True})
app.config['DATABASE'] = settings.DATABASE_CONFIG
db = Database(app)
from redash.authentication import setup_authentication
auth = setup_authentication(app)
@api.representation('application/json')
def json_representation(data, code, headers=None):
resp = make_response(json.dumps(data, cls=utils.JSONEncoder), code)
resp.headers.extend(headers or {})
return resp
redis_url = urlparse.urlparse(settings.REDIS_URL)
if redis_url.path:
redis_db = redis_url.path[1]
else:
redis_db = 0
redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password)
from redash import data
data_manager = data.Manager(redis_connection, db)
from redash import controllers | // ... existing code ...
redis_url = urlparse.urlparse(settings.REDIS_URL)
if redis_url.path:
redis_db = redis_url.path[1]
else:
redis_db = 0
redis_connection = redis.StrictRedis(host=redis_url.hostname, port=redis_url.port, db=redis_db, password=redis_url.password)
from redash import data
// ... rest of the code ... |
c4e1f1c147783a4a735dd943d5d7491302de300e | csunplugged/config/urls.py | csunplugged/config/urls.py | from django.conf.urls import include, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
from django.conf import settings
from django.conf.urls.static import static
urlpatterns = i18n_patterns(
url(r'', include('general.urls', namespace='general')),
url(r'^topics/', include('topics.urls', namespace='topics')),
url(r'^resources/', include('resources.urls', namespace='resources')),
url(r'^admin/', include(admin.site.urls)),
)
# ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT)
| from django.conf.urls import include, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
urlpatterns = i18n_patterns(
url(r'', include('general.urls', namespace='general')),
url(r'^topics/', include('topics.urls', namespace='topics')),
url(r'^resources/', include('resources.urls', namespace='resources')),
url(r'^admin/', include(admin.site.urls)),
)
| Remove unused static URL pathing | Remove unused static URL pathing
| Python | mit | uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged | from django.conf.urls import include, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
- from django.conf import settings
- from django.conf.urls.static import static
urlpatterns = i18n_patterns(
url(r'', include('general.urls', namespace='general')),
url(r'^topics/', include('topics.urls', namespace='topics')),
url(r'^resources/', include('resources.urls', namespace='resources')),
url(r'^admin/', include(admin.site.urls)),
)
- # ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT)
| Remove unused static URL pathing | ## Code Before:
from django.conf.urls import include, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
from django.conf import settings
from django.conf.urls.static import static
urlpatterns = i18n_patterns(
url(r'', include('general.urls', namespace='general')),
url(r'^topics/', include('topics.urls', namespace='topics')),
url(r'^resources/', include('resources.urls', namespace='resources')),
url(r'^admin/', include(admin.site.urls)),
)
# ] + static(settings.STATIC_URL, documnet_root=settings.STATIC_ROOT)
## Instruction:
Remove unused static URL pathing
## Code After:
from django.conf.urls import include, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
urlpatterns = i18n_patterns(
url(r'', include('general.urls', namespace='general')),
url(r'^topics/', include('topics.urls', namespace='topics')),
url(r'^resources/', include('resources.urls', namespace='resources')),
url(r'^admin/', include(admin.site.urls)),
)
| ...
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
urlpatterns = i18n_patterns(
...
url(r'^admin/', include(admin.site.urls)),
)
... |
252ffda53d494403133fdb1986c92422264406d8 | tests_app/tests/unit/serializers/models.py | tests_app/tests/unit/serializers/models.py | import os
from django.db import models
from django.conf import settings
class UserModel(models.Model):
name = models.CharField(max_length=20)
upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers')
class CommentModel(models.Model):
user = models.ForeignKey(
UserModel,
related_name='comments',
on_delete=models.CASCADE,
)
users_liked = models.ManyToManyField(UserModel, blank=True, null=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
attachment = models.FileField(
upload_to=upload_to, blank=True, null=True, max_length=500)
hidden_text = models.CharField(max_length=200, blank=True, null=True)
| import os
from django.db import models
from django.conf import settings
class UserModel(models.Model):
name = models.CharField(max_length=20)
upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers')
class CommentModel(models.Model):
user = models.ForeignKey(
UserModel,
related_name='comments',
on_delete=models.CASCADE,
)
users_liked = models.ManyToManyField(UserModel, blank=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
attachment = models.FileField(
upload_to=upload_to, blank=True, null=True, max_length=500)
hidden_text = models.CharField(max_length=200, blank=True, null=True)
| Fix CommentModel m2m null warning | Fix CommentModel m2m null warning
| Python | mit | chibisov/drf-extensions | import os
from django.db import models
from django.conf import settings
class UserModel(models.Model):
name = models.CharField(max_length=20)
upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers')
class CommentModel(models.Model):
user = models.ForeignKey(
UserModel,
related_name='comments',
on_delete=models.CASCADE,
)
- users_liked = models.ManyToManyField(UserModel, blank=True, null=True)
+ users_liked = models.ManyToManyField(UserModel, blank=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
attachment = models.FileField(
upload_to=upload_to, blank=True, null=True, max_length=500)
hidden_text = models.CharField(max_length=200, blank=True, null=True)
| Fix CommentModel m2m null warning | ## Code Before:
import os
from django.db import models
from django.conf import settings
class UserModel(models.Model):
name = models.CharField(max_length=20)
upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers')
class CommentModel(models.Model):
user = models.ForeignKey(
UserModel,
related_name='comments',
on_delete=models.CASCADE,
)
users_liked = models.ManyToManyField(UserModel, blank=True, null=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
attachment = models.FileField(
upload_to=upload_to, blank=True, null=True, max_length=500)
hidden_text = models.CharField(max_length=200, blank=True, null=True)
## Instruction:
Fix CommentModel m2m null warning
## Code After:
import os
from django.db import models
from django.conf import settings
class UserModel(models.Model):
name = models.CharField(max_length=20)
upload_to = os.path.join(settings.FILE_STORAGE_DIR, 'test_serializers')
class CommentModel(models.Model):
user = models.ForeignKey(
UserModel,
related_name='comments',
on_delete=models.CASCADE,
)
users_liked = models.ManyToManyField(UserModel, blank=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
attachment = models.FileField(
upload_to=upload_to, blank=True, null=True, max_length=500)
hidden_text = models.CharField(max_length=200, blank=True, null=True)
| ...
on_delete=models.CASCADE,
)
users_liked = models.ManyToManyField(UserModel, blank=True)
title = models.CharField(max_length=20)
text = models.CharField(max_length=200)
... |
20147b8b8a80ef8ab202d916bf1cdfb67d4753d3 | SelfTests.py | SelfTests.py | import os
import unittest
from Logger import Logger
class TestLogger(unittest.TestCase):
def test_file_handling(self):
testLog = Logger("testLog")
## Check if program can create and open file
self.assertTrue(testLog.opened)
returns = testLog.close()
## Check if logger correctly signs bool OPENED and returns
## 0 as succes.
self.assertFalse(testLog.opened)
self.assertEqual(returns,0)
returns = testLog.close()
## Check if logger returns 1 when trying to close already
## closed file
self.assertEqual(returns,1)
## Do cleanup:
os.remove(testLog.name)
def test_logging(self):
testLog = Logger("testLog")
testLog.save_line("TestLine")
testLog.close()
logfile = open(testLog.name)
content = logfile.read()
logfile.close()
saved = content.split(" : ")
self.assertEqual(saved[1],"TestLine")
## cleanup
os.remove(testLog.name)
if __name__ == '__main__':
unittest.main()
| import os
import unittest
from Logger import Logger
class TestLogger(unittest.TestCase):
def test_file_handling(self):
testLog = Logger("testLog")
## Check if program can create and open file
self.assertTrue(testLog.opened)
returns = testLog.close()
## Check if logger correctly signs bool OPENED and returns
## 0 as succes.
self.assertFalse(testLog.opened)
self.assertEqual(returns,0)
returns = testLog.close()
## Check if logger returns 1 when trying to close already
## closed file
self.assertEqual(returns,1)
## Do cleanup:
os.remove(testLog.name)
def test_logging(self):
testLog = Logger("testLog")
testPhrase = "TestLine\r\n"
testLog.save_line(testPhrase)
testLog.close()
logfile = open(testLog.name)
content = logfile.read()
logfile.close()
saved = content.split(" : ")
## Check if saved data corresponds
self.assertEqual(saved[1],testPhrase)
## cleanup
os.remove(testLog.name)
if __name__ == '__main__':
unittest.main()
| Test of logger is testing an testPhrase instead of two manually writen strings | Test of logger is testing an testPhrase instead of two manually writen strings
Signed-off-by: TeaPackCZ <[email protected]>
| Python | mit | TeaPackCZ/RobotZed,TeaPackCZ/RobotZed | import os
import unittest
from Logger import Logger
class TestLogger(unittest.TestCase):
def test_file_handling(self):
testLog = Logger("testLog")
## Check if program can create and open file
self.assertTrue(testLog.opened)
returns = testLog.close()
## Check if logger correctly signs bool OPENED and returns
## 0 as succes.
self.assertFalse(testLog.opened)
self.assertEqual(returns,0)
returns = testLog.close()
## Check if logger returns 1 when trying to close already
## closed file
self.assertEqual(returns,1)
## Do cleanup:
os.remove(testLog.name)
def test_logging(self):
testLog = Logger("testLog")
+ testPhrase = "TestLine\r\n"
- testLog.save_line("TestLine")
+ testLog.save_line(testPhrase)
testLog.close()
logfile = open(testLog.name)
content = logfile.read()
logfile.close()
saved = content.split(" : ")
+ ## Check if saved data corresponds
- self.assertEqual(saved[1],"TestLine")
+ self.assertEqual(saved[1],testPhrase)
## cleanup
os.remove(testLog.name)
if __name__ == '__main__':
unittest.main()
| Test of logger is testing an testPhrase instead of two manually writen strings | ## Code Before:
import os
import unittest
from Logger import Logger
class TestLogger(unittest.TestCase):
def test_file_handling(self):
testLog = Logger("testLog")
## Check if program can create and open file
self.assertTrue(testLog.opened)
returns = testLog.close()
## Check if logger correctly signs bool OPENED and returns
## 0 as succes.
self.assertFalse(testLog.opened)
self.assertEqual(returns,0)
returns = testLog.close()
## Check if logger returns 1 when trying to close already
## closed file
self.assertEqual(returns,1)
## Do cleanup:
os.remove(testLog.name)
def test_logging(self):
testLog = Logger("testLog")
testLog.save_line("TestLine")
testLog.close()
logfile = open(testLog.name)
content = logfile.read()
logfile.close()
saved = content.split(" : ")
self.assertEqual(saved[1],"TestLine")
## cleanup
os.remove(testLog.name)
if __name__ == '__main__':
unittest.main()
## Instruction:
Test of logger is testing an testPhrase instead of two manually writen strings
## Code After:
import os
import unittest
from Logger import Logger
class TestLogger(unittest.TestCase):
def test_file_handling(self):
testLog = Logger("testLog")
## Check if program can create and open file
self.assertTrue(testLog.opened)
returns = testLog.close()
## Check if logger correctly signs bool OPENED and returns
## 0 as succes.
self.assertFalse(testLog.opened)
self.assertEqual(returns,0)
returns = testLog.close()
## Check if logger returns 1 when trying to close already
## closed file
self.assertEqual(returns,1)
## Do cleanup:
os.remove(testLog.name)
def test_logging(self):
testLog = Logger("testLog")
testPhrase = "TestLine\r\n"
testLog.save_line(testPhrase)
testLog.close()
logfile = open(testLog.name)
content = logfile.read()
logfile.close()
saved = content.split(" : ")
## Check if saved data corresponds
self.assertEqual(saved[1],testPhrase)
## cleanup
os.remove(testLog.name)
if __name__ == '__main__':
unittest.main()
| # ... existing code ...
def test_logging(self):
testLog = Logger("testLog")
testPhrase = "TestLine\r\n"
testLog.save_line(testPhrase)
testLog.close()
logfile = open(testLog.name)
# ... modified code ...
logfile.close()
saved = content.split(" : ")
## Check if saved data corresponds
self.assertEqual(saved[1],testPhrase)
## cleanup
os.remove(testLog.name)
# ... rest of the code ... |
bb0b72333b715956740373c3ba80a8193b99a8cc | app/services/updater_service.py | app/services/updater_service.py | from app.system.updater import check_updates, do_upgrade, run_ansible
from app.views import SimpleBackgroundView
from .base import BaseService, BlockingServiceStart
class UpdaterService(BaseService, BlockingServiceStart):
def __init__(self, observer=None):
super().__init__(observer=observer)
self._view = SimpleBackgroundView("Checking for updates.")
def on_service_start(self):
values = check_updates()
for val in values:
self._view.args["subtitle"] = "Working with: " + str(val)
do_upgrade([val])
if values:
run_ansible()
def view(self):
return self._view
| from app.system.updater import check_updates, do_upgrade, run_ansible
from app.views import SimpleBackgroundView
from .base import BaseService, BlockingServiceStart
class UpdaterService(BaseService, BlockingServiceStart):
def __init__(self, observer=None):
super().__init__(observer=observer)
self._view = SimpleBackgroundView("Checking for updates.")
def on_service_start(self):
values = check_updates()
for val in values:
self._view.args["subtitle"] = "Working with: " + str(val)
do_upgrade([val])
if values:
self._view.args["subtitle"] = "Finishing up..."
run_ansible()
def view(self):
return self._view
| Add message before running ansible. | Add message before running ansible.
| Python | mit | supersaiyanmode/HomePiServer,supersaiyanmode/HomePiServer,supersaiyanmode/HomePiServer | from app.system.updater import check_updates, do_upgrade, run_ansible
from app.views import SimpleBackgroundView
from .base import BaseService, BlockingServiceStart
class UpdaterService(BaseService, BlockingServiceStart):
def __init__(self, observer=None):
super().__init__(observer=observer)
self._view = SimpleBackgroundView("Checking for updates.")
def on_service_start(self):
values = check_updates()
for val in values:
self._view.args["subtitle"] = "Working with: " + str(val)
do_upgrade([val])
if values:
+ self._view.args["subtitle"] = "Finishing up..."
run_ansible()
def view(self):
return self._view
| Add message before running ansible. | ## Code Before:
from app.system.updater import check_updates, do_upgrade, run_ansible
from app.views import SimpleBackgroundView
from .base import BaseService, BlockingServiceStart
class UpdaterService(BaseService, BlockingServiceStart):
def __init__(self, observer=None):
super().__init__(observer=observer)
self._view = SimpleBackgroundView("Checking for updates.")
def on_service_start(self):
values = check_updates()
for val in values:
self._view.args["subtitle"] = "Working with: " + str(val)
do_upgrade([val])
if values:
run_ansible()
def view(self):
return self._view
## Instruction:
Add message before running ansible.
## Code After:
from app.system.updater import check_updates, do_upgrade, run_ansible
from app.views import SimpleBackgroundView
from .base import BaseService, BlockingServiceStart
class UpdaterService(BaseService, BlockingServiceStart):
def __init__(self, observer=None):
super().__init__(observer=observer)
self._view = SimpleBackgroundView("Checking for updates.")
def on_service_start(self):
values = check_updates()
for val in values:
self._view.args["subtitle"] = "Working with: " + str(val)
do_upgrade([val])
if values:
self._view.args["subtitle"] = "Finishing up..."
run_ansible()
def view(self):
return self._view
| ...
do_upgrade([val])
if values:
self._view.args["subtitle"] = "Finishing up..."
run_ansible()
... |
866026a5d2f89a8ac76a726720e4fbe812c94eb4 | ds/providers/shell.py | ds/providers/shell.py | from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format({
'environment': task.environment,
'sha': task.sha,
'ref': task.ref,
'task': task.name,
})
return workspace.run(command)
| from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format(
environment=task.environment,
sha=task.sha,
ref=task.ref,
task=task.name,
)
return workspace.run(command)
| Fix arg passing to command | Fix arg passing to command
| Python | apache-2.0 | rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,getsentry/freight,klynton/freight,getsentry/freight,klynton/freight,rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,rshk/freight,jkimbo/freight,jkimbo/freight,rshk/freight,getsentry/freight | from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
- command = task.provider_config['command'].format({
+ command = task.provider_config['command'].format(
- 'environment': task.environment,
+ environment=task.environment,
- 'sha': task.sha,
+ sha=task.sha,
- 'ref': task.ref,
+ ref=task.ref,
- 'task': task.name,
+ task=task.name,
- })
+ )
return workspace.run(command)
| Fix arg passing to command | ## Code Before:
from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format({
'environment': task.environment,
'sha': task.sha,
'ref': task.ref,
'task': task.name,
})
return workspace.run(command)
## Instruction:
Fix arg passing to command
## Code After:
from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format(
environment=task.environment,
sha=task.sha,
ref=task.ref,
task=task.name,
)
return workspace.run(command)
| ...
def execute(self, workspace, task):
command = task.provider_config['command'].format(
environment=task.environment,
sha=task.sha,
ref=task.ref,
task=task.name,
)
return workspace.run(command)
... |
59ec54bbe49013826d2c15ce2162c2e0e335bd57 | modules/module_urlsize.py | modules/module_urlsize.py | """Warns about large files"""
def handle_url(bot, user, channel, url, msg):
if channel == "#wow": return
# inform about large files (over 5MB)
size = getUrl(url).getSize()
contentType = getUrl(url).getHeaders()['content-type']
if not size: return
size = size / 1024
if size > 5:
bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
| """Warns about large files"""
def handle_url(bot, user, channel, url, msg):
if channel == "#wow": return
# inform about large files (over 5MB)
size = getUrl(url).getSize()
headers = getUrl(url).getHeaders()['content-type']
if 'content-type' in headers:
contentType = headers['content-type']
else:
contentType = "Unknown"
if not size: return
size = size / 1024
if size > 5:
bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
| Handle cases where the server doesn't return content-type | Handle cases where the server doesn't return content-type
git-svn-id: 056f9092885898c4775d98c479d2d33d00273e45@120 dda364a1-ef19-0410-af65-756c83048fb2
| Python | bsd-3-clause | rnyberg/pyfibot,lepinkainen/pyfibot,lepinkainen/pyfibot,rnyberg/pyfibot,EArmour/pyfibot,nigeljonez/newpyfibot,aapa/pyfibot,huqa/pyfibot,huqa/pyfibot,aapa/pyfibot,EArmour/pyfibot | """Warns about large files"""
def handle_url(bot, user, channel, url, msg):
if channel == "#wow": return
# inform about large files (over 5MB)
size = getUrl(url).getSize()
- contentType = getUrl(url).getHeaders()['content-type']
+ headers = getUrl(url).getHeaders()['content-type']
+ if 'content-type' in headers:
+ contentType = headers['content-type']
+ else:
+ contentType = "Unknown"
if not size: return
size = size / 1024
if size > 5:
bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
| Handle cases where the server doesn't return content-type | ## Code Before:
"""Warns about large files"""
def handle_url(bot, user, channel, url, msg):
if channel == "#wow": return
# inform about large files (over 5MB)
size = getUrl(url).getSize()
contentType = getUrl(url).getHeaders()['content-type']
if not size: return
size = size / 1024
if size > 5:
bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
## Instruction:
Handle cases where the server doesn't return content-type
## Code After:
"""Warns about large files"""
def handle_url(bot, user, channel, url, msg):
if channel == "#wow": return
# inform about large files (over 5MB)
size = getUrl(url).getSize()
headers = getUrl(url).getHeaders()['content-type']
if 'content-type' in headers:
contentType = headers['content-type']
else:
contentType = "Unknown"
if not size: return
size = size / 1024
if size > 5:
bot.say(channel, "File size: %s MB - Content-Type: %s" % (size, contentType))
| # ... existing code ...
# inform about large files (over 5MB)
size = getUrl(url).getSize()
headers = getUrl(url).getHeaders()['content-type']
if 'content-type' in headers:
contentType = headers['content-type']
else:
contentType = "Unknown"
if not size: return
# ... rest of the code ... |
056d82002c133736a800b08bd071b71c9f5615f8 | ci/generate_pipeline_yml.py | ci/generate_pipeline_yml.py |
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
| Update TAS versions we test against | Update TAS versions we test against
| Python | apache-2.0 | cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator |
import os
from jinja2 import Template
- clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
+ clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
| Update TAS versions we test against | ## Code Before:
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
## Instruction:
Update TAS versions we test against
## Code After:
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
| // ... existing code ...
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
// ... rest of the code ... |
bbfdbc4b5b6a35105a65910a878be85040cf5263 | VMEncryption/main/oscrypto/encryptstates/OSEncryptionState.py | VMEncryption/main/oscrypto/encryptstates/OSEncryptionState.py |
from collections import namedtuple
class OSEncryptionState(object):
def __init__(self, context):
super(OSEncryptionState, self).__init__()
self.state_executed = False
def enter(self):
assert 0, "implement enter"
def should_exit(self):
assert 0, "implement should_exit"
OSEncryptionStateContext = namedtuple('OSEncryptionStateContext',
['hutil',
'distro_patcher',
'logger',
'encryption_environment'])
|
from collections import namedtuple
class OSEncryptionState(object):
def __init__(self, context):
super(OSEncryptionState, self).__init__()
def enter(self):
assert 0, "implement enter"
def should_exit(self):
assert 0, "implement should_exit"
OSEncryptionStateContext = namedtuple('OSEncryptionStateContext',
['hutil',
'distro_patcher',
'logger',
'encryption_environment'])
| Remove var declaration from abstract base class | Remove var declaration from abstract base class
| Python | apache-2.0 | soumyanishan/azure-linux-extensions,bpramod/azure-linux-extensions,Azure/azure-linux-extensions,Azure/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,Azure/azure-linux-extensions,andyliuliming/azure-linux-extensions,andyliuliming/azure-linux-extensions,jasonzio/azure-linux-extensions,krkhan/azure-linux-extensions,varunkumta/azure-linux-extensions,Azure/azure-linux-extensions,jasonzio/azure-linux-extensions,krkhan/azure-linux-extensions,bpramod/azure-linux-extensions,krkhan/azure-linux-extensions,jasonzio/azure-linux-extensions,soumyanishan/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,soumyanishan/azure-linux-extensions,soumyanishan/azure-linux-extensions,varunkumta/azure-linux-extensions,varunkumta/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,varunkumta/azure-linux-extensions,andyliuliming/azure-linux-extensions,andyliuliming/azure-linux-extensions,bpramod/azure-linux-extensions,bpramod/azure-linux-extensions,vityagi/azure-linux-extensions,vityagi/azure-linux-extensions,krkhan/azure-linux-extensions,vityagi/azure-linux-extensions,jasonzio/azure-linux-extensions,vityagi/azure-linux-extensions,Azure/azure-linux-extensions,bpramod/azure-linux-extensions,soumyanishan/azure-linux-extensions |
from collections import namedtuple
class OSEncryptionState(object):
def __init__(self, context):
super(OSEncryptionState, self).__init__()
-
- self.state_executed = False
def enter(self):
assert 0, "implement enter"
def should_exit(self):
assert 0, "implement should_exit"
OSEncryptionStateContext = namedtuple('OSEncryptionStateContext',
['hutil',
'distro_patcher',
'logger',
'encryption_environment'])
| Remove var declaration from abstract base class | ## Code Before:
from collections import namedtuple
class OSEncryptionState(object):
def __init__(self, context):
super(OSEncryptionState, self).__init__()
self.state_executed = False
def enter(self):
assert 0, "implement enter"
def should_exit(self):
assert 0, "implement should_exit"
OSEncryptionStateContext = namedtuple('OSEncryptionStateContext',
['hutil',
'distro_patcher',
'logger',
'encryption_environment'])
## Instruction:
Remove var declaration from abstract base class
## Code After:
from collections import namedtuple
class OSEncryptionState(object):
def __init__(self, context):
super(OSEncryptionState, self).__init__()
def enter(self):
assert 0, "implement enter"
def should_exit(self):
assert 0, "implement should_exit"
OSEncryptionStateContext = namedtuple('OSEncryptionStateContext',
['hutil',
'distro_patcher',
'logger',
'encryption_environment'])
| # ... existing code ...
def __init__(self, context):
super(OSEncryptionState, self).__init__()
def enter(self):
# ... rest of the code ... |
5b3d38821517f10f9b9da31f28af19e7302de954 | dimod/reference/composites/structure.py | dimod/reference/composites/structure.py | from dimod.core.sampler import Sampler
from dimod.core.composite import Composite
from dimod.core.structured import Structured
from dimod.decorators import bqm_structured
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
todo
"""
def __init__(self, sampler, nodelist, edgelist):
Sampler.__init__(self)
Composite.__init__(self, sampler)
Structured.__init__(self, nodelist, edgelist)
@bqm_structured
def sample(self, bqm, **sample_kwargs):
return self.child.sample(bqm, **sample_kwargs)
| from dimod.core.sampler import Sampler
from dimod.core.composite import Composite
from dimod.core.structured import Structured
from dimod.decorators import bqm_structured
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
"""
# we will override these in the __init__, but because they are abstract properties we need to
# signal that we are overriding them
edgelist = None
nodelist = None
children = None
def __init__(self, sampler, nodelist, edgelist):
self.children = [sampler]
self.nodelist = nodelist
self.edgelist = edgelist
@property
def parameters(self):
return self.child.parameters
@property
def properties(self):
return self.child.properties
@bqm_structured
def sample(self, bqm, **sample_kwargs):
return self.child.sample(bqm, **sample_kwargs)
| Update Structure composite to use the new abc | Update Structure composite to use the new abc
| Python | apache-2.0 | oneklc/dimod,oneklc/dimod | from dimod.core.sampler import Sampler
from dimod.core.composite import Composite
from dimod.core.structured import Structured
from dimod.decorators import bqm_structured
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
+ """
+ # we will override these in the __init__, but because they are abstract properties we need to
+ # signal that we are overriding them
+ edgelist = None
+ nodelist = None
+ children = None
- todo
- """
def __init__(self, sampler, nodelist, edgelist):
- Sampler.__init__(self)
- Composite.__init__(self, sampler)
- Structured.__init__(self, nodelist, edgelist)
+ self.children = [sampler]
+ self.nodelist = nodelist
+ self.edgelist = edgelist
+
+ @property
+ def parameters(self):
+ return self.child.parameters
+
+ @property
+ def properties(self):
+ return self.child.properties
@bqm_structured
def sample(self, bqm, **sample_kwargs):
return self.child.sample(bqm, **sample_kwargs)
| Update Structure composite to use the new abc | ## Code Before:
from dimod.core.sampler import Sampler
from dimod.core.composite import Composite
from dimod.core.structured import Structured
from dimod.decorators import bqm_structured
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
todo
"""
def __init__(self, sampler, nodelist, edgelist):
Sampler.__init__(self)
Composite.__init__(self, sampler)
Structured.__init__(self, nodelist, edgelist)
@bqm_structured
def sample(self, bqm, **sample_kwargs):
return self.child.sample(bqm, **sample_kwargs)
## Instruction:
Update Structure composite to use the new abc
## Code After:
from dimod.core.sampler import Sampler
from dimod.core.composite import Composite
from dimod.core.structured import Structured
from dimod.decorators import bqm_structured
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
"""
# we will override these in the __init__, but because they are abstract properties we need to
# signal that we are overriding them
edgelist = None
nodelist = None
children = None
def __init__(self, sampler, nodelist, edgelist):
self.children = [sampler]
self.nodelist = nodelist
self.edgelist = edgelist
@property
def parameters(self):
return self.child.parameters
@property
def properties(self):
return self.child.properties
@bqm_structured
def sample(self, bqm, **sample_kwargs):
return self.child.sample(bqm, **sample_kwargs)
| // ... existing code ...
class StructureComposite(Sampler, Composite, Structured):
"""Creates a structured composed sampler from an unstructured sampler.
"""
# we will override these in the __init__, but because they are abstract properties we need to
# signal that we are overriding them
edgelist = None
nodelist = None
children = None
def __init__(self, sampler, nodelist, edgelist):
self.children = [sampler]
self.nodelist = nodelist
self.edgelist = edgelist
@property
def parameters(self):
return self.child.parameters
@property
def properties(self):
return self.child.properties
@bqm_structured
// ... rest of the code ... |
bb578d4237ccaf16fe5c38842cc100cdbefc0119 | senlin/tests/functional/drivers/openstack/__init__.py | senlin/tests/functional/drivers/openstack/__init__.py |
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
from senlin.tests.functional.drivers.openstack import nova_v2
# Currently, only fake nova_v2 driver is supported
def compute(params):
return nova_v2.NovaClient(params)
def loadbalancing(params):
return lbaas.LoadBalancerDriver(params)
def network(params):
return neutron_v2.NeutronClient(params)
def orchestration(params):
return heat_v1.HeatClient(params)
def telemetry(params):
return ceilometer_v2.CeilometerClient(params)
|
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
from senlin.drivers.openstack import keystone_v3
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
from senlin.tests.functional.drivers.openstack import nova_v2
# Currently, only fake nova_v2 driver is supported
compute = nova_v2.NovaClient
identity = keystone_v3.KeystoneClient
loadbalancing = lbaas.LoadBalancerDriver
network = neutron_v2.NeutronClient
orchestration = heat_v1.HeatClient
telemetry = ceilometer_v2.CeilometerClient
| Add keystone driver plugin for functional test | Add keystone driver plugin for functional test
This patch adds keystone driver plugin for functional test.
Change-Id: Iefa9c1b8956854ae75f672627aa3d2f9f7d22c0e
| Python | apache-2.0 | openstack/senlin,stackforge/senlin,tengqm/senlin-container,stackforge/senlin,openstack/senlin,tengqm/senlin-container,Alzon/senlin,Alzon/senlin,openstack/senlin |
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
+ from senlin.drivers.openstack import keystone_v3
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
from senlin.tests.functional.drivers.openstack import nova_v2
# Currently, only fake nova_v2 driver is supported
- def compute(params):
- return nova_v2.NovaClient(params)
+ compute = nova_v2.NovaClient
+ identity = keystone_v3.KeystoneClient
+ loadbalancing = lbaas.LoadBalancerDriver
+ network = neutron_v2.NeutronClient
+ orchestration = heat_v1.HeatClient
+ telemetry = ceilometer_v2.CeilometerClient
-
- def loadbalancing(params):
- return lbaas.LoadBalancerDriver(params)
-
-
- def network(params):
- return neutron_v2.NeutronClient(params)
-
-
- def orchestration(params):
- return heat_v1.HeatClient(params)
-
-
- def telemetry(params):
- return ceilometer_v2.CeilometerClient(params)
- | Add keystone driver plugin for functional test | ## Code Before:
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
from senlin.tests.functional.drivers.openstack import nova_v2
# Currently, only fake nova_v2 driver is supported
def compute(params):
return nova_v2.NovaClient(params)
def loadbalancing(params):
return lbaas.LoadBalancerDriver(params)
def network(params):
return neutron_v2.NeutronClient(params)
def orchestration(params):
return heat_v1.HeatClient(params)
def telemetry(params):
return ceilometer_v2.CeilometerClient(params)
## Instruction:
Add keystone driver plugin for functional test
## Code After:
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
from senlin.drivers.openstack import keystone_v3
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
from senlin.tests.functional.drivers.openstack import nova_v2
# Currently, only fake nova_v2 driver is supported
compute = nova_v2.NovaClient
identity = keystone_v3.KeystoneClient
loadbalancing = lbaas.LoadBalancerDriver
network = neutron_v2.NeutronClient
orchestration = heat_v1.HeatClient
telemetry = ceilometer_v2.CeilometerClient
| // ... existing code ...
from senlin.drivers.openstack import ceilometer_v2
from senlin.drivers.openstack import heat_v1
from senlin.drivers.openstack import keystone_v3
from senlin.drivers.openstack import lbaas
from senlin.drivers.openstack import neutron_v2
// ... modified code ...
# Currently, only fake nova_v2 driver is supported
compute = nova_v2.NovaClient
identity = keystone_v3.KeystoneClient
loadbalancing = lbaas.LoadBalancerDriver
network = neutron_v2.NeutronClient
orchestration = heat_v1.HeatClient
telemetry = ceilometer_v2.CeilometerClient
// ... rest of the code ... |
cc6ce477550152135eed5a9e35bca8144be10111 | groupmestats/plotly_helpers.py | groupmestats/plotly_helpers.py | import plotly
def try_saving_plotly_figure(figure, filename):
try:
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
if 'The response from plotly could not be translated.'in str(e):
print("Failed to save plotly figure. <home>/.plotly/.credentials"
" might not be configured correctly? "
"Or you may have hit your plotly account's rate limit"
" (http://help.plot.ly/api-rate-limits/)")
else:
raise
# A green bar with slightly darker green line
marker = dict(
color='#4BB541',
line=dict(
color='#3A9931',
width=1.5,
)
)
| import plotly
def try_saving_plotly_figure(figure, filename):
try:
print("Saving plot to '%s'" % filename)
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
if 'The response from plotly could not be translated.'in str(e):
print("Failed to save plotly figure. <home>/.plotly/.credentials"
" might not be configured correctly? "
"Or you may have hit your plotly account's rate limit"
" (http://help.plot.ly/api-rate-limits/)")
else:
raise
# A green bar with slightly darker green line
marker = dict(
color='#4BB541',
line=dict(
color='#3A9931',
width=1.5,
)
)
| Print when saving plot to file | Print when saving plot to file
| Python | mit | kjteske/groupmestats,kjteske/groupmestats | import plotly
def try_saving_plotly_figure(figure, filename):
try:
+ print("Saving plot to '%s'" % filename)
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
if 'The response from plotly could not be translated.'in str(e):
print("Failed to save plotly figure. <home>/.plotly/.credentials"
" might not be configured correctly? "
"Or you may have hit your plotly account's rate limit"
" (http://help.plot.ly/api-rate-limits/)")
else:
raise
# A green bar with slightly darker green line
marker = dict(
color='#4BB541',
line=dict(
color='#3A9931',
width=1.5,
)
)
| Print when saving plot to file | ## Code Before:
import plotly
def try_saving_plotly_figure(figure, filename):
try:
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
if 'The response from plotly could not be translated.'in str(e):
print("Failed to save plotly figure. <home>/.plotly/.credentials"
" might not be configured correctly? "
"Or you may have hit your plotly account's rate limit"
" (http://help.plot.ly/api-rate-limits/)")
else:
raise
# A green bar with slightly darker green line
marker = dict(
color='#4BB541',
line=dict(
color='#3A9931',
width=1.5,
)
)
## Instruction:
Print when saving plot to file
## Code After:
import plotly
def try_saving_plotly_figure(figure, filename):
try:
print("Saving plot to '%s'" % filename)
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
if 'The response from plotly could not be translated.'in str(e):
print("Failed to save plotly figure. <home>/.plotly/.credentials"
" might not be configured correctly? "
"Or you may have hit your plotly account's rate limit"
" (http://help.plot.ly/api-rate-limits/)")
else:
raise
# A green bar with slightly darker green line
marker = dict(
color='#4BB541',
line=dict(
color='#3A9931',
width=1.5,
)
)
| // ... existing code ...
def try_saving_plotly_figure(figure, filename):
try:
print("Saving plot to '%s'" % filename)
plotly.plotly.image.save_as(figure, filename)
except plotly.exceptions.PlotlyError as e:
// ... rest of the code ... |
a5ff4c247030559c83a06976fcda062c0c42d810 | django_fixmystreet/fixmystreet/tests/__init__.py | django_fixmystreet/fixmystreet/tests/__init__.py | import shutil
import os
from django.core.files.storage import default_storage
from django.test import TestCase
class SampleFilesTestCase(TestCase):
fixtures = ['sample']
@classmethod
def setUpClass(cls):
default_storage.location = 'media' # force using source media folder to avoid real data erasing
# @classmethod
# def setUpClass(cls):
# shutil.copytree('media', 'media-tmp')
# default_storage.location = 'media-tmp'
#
# @classmethod
# def tearDownClass(self):
# shutil.rmtree('media-tmp')
def _fixture_setup(self):
if os.path.exists('media/photos'):
shutil.rmtree('media/photos')
shutil.copytree('media/photos-sample', 'media/photos')
super(SampleFilesTestCase, self)._fixture_setup()
def tearDown(self):
shutil.rmtree('media/photos')
from django_fixmystreet.fixmystreet.tests.views import *
from django_fixmystreet.fixmystreet.tests.reports import *
from django_fixmystreet.fixmystreet.tests.users import *
from django_fixmystreet.fixmystreet.tests.organisation_entity import *
from django_fixmystreet.fixmystreet.tests.mail import *
# from django_fixmystreet.fixmystreet.tests.api import *
| import shutil
import os
from django.core.files.storage import default_storage
from django.test import TestCase
class SampleFilesTestCase(TestCase):
fixtures = ['sample']
@classmethod
def setUpClass(cls):
default_storage.location = 'media' # force using source media folder to avoid real data erasing
# @classmethod
# def setUpClass(cls):
# shutil.copytree('media', 'media-tmp')
# default_storage.location = 'media-tmp'
#
# @classmethod
# def tearDownClass(self):
# shutil.rmtree('media-tmp')
def _fixture_setup(self):
if os.path.exists('media/photos'):
shutil.rmtree('media/photos')
shutil.copytree('media/photos-sample', 'media/photos')
super(SampleFilesTestCase, self)._fixture_setup()
from django_fixmystreet.fixmystreet.tests.views import *
from django_fixmystreet.fixmystreet.tests.reports import *
from django_fixmystreet.fixmystreet.tests.users import *
from django_fixmystreet.fixmystreet.tests.organisation_entity import *
from django_fixmystreet.fixmystreet.tests.mail import *
# from django_fixmystreet.fixmystreet.tests.api import *
| Fix unit test fixtures files | Fix unit test fixtures files
| Python | agpl-3.0 | IMIO/django-fixmystreet,IMIO/django-fixmystreet,IMIO/django-fixmystreet,IMIO/django-fixmystreet | import shutil
import os
from django.core.files.storage import default_storage
from django.test import TestCase
class SampleFilesTestCase(TestCase):
fixtures = ['sample']
@classmethod
def setUpClass(cls):
default_storage.location = 'media' # force using source media folder to avoid real data erasing
# @classmethod
# def setUpClass(cls):
# shutil.copytree('media', 'media-tmp')
# default_storage.location = 'media-tmp'
#
# @classmethod
# def tearDownClass(self):
# shutil.rmtree('media-tmp')
def _fixture_setup(self):
if os.path.exists('media/photos'):
shutil.rmtree('media/photos')
shutil.copytree('media/photos-sample', 'media/photos')
super(SampleFilesTestCase, self)._fixture_setup()
- def tearDown(self):
- shutil.rmtree('media/photos')
from django_fixmystreet.fixmystreet.tests.views import *
from django_fixmystreet.fixmystreet.tests.reports import *
from django_fixmystreet.fixmystreet.tests.users import *
from django_fixmystreet.fixmystreet.tests.organisation_entity import *
from django_fixmystreet.fixmystreet.tests.mail import *
# from django_fixmystreet.fixmystreet.tests.api import *
| Fix unit test fixtures files | ## Code Before:
import shutil
import os
from django.core.files.storage import default_storage
from django.test import TestCase
class SampleFilesTestCase(TestCase):
fixtures = ['sample']
@classmethod
def setUpClass(cls):
default_storage.location = 'media' # force using source media folder to avoid real data erasing
# @classmethod
# def setUpClass(cls):
# shutil.copytree('media', 'media-tmp')
# default_storage.location = 'media-tmp'
#
# @classmethod
# def tearDownClass(self):
# shutil.rmtree('media-tmp')
def _fixture_setup(self):
if os.path.exists('media/photos'):
shutil.rmtree('media/photos')
shutil.copytree('media/photos-sample', 'media/photos')
super(SampleFilesTestCase, self)._fixture_setup()
def tearDown(self):
shutil.rmtree('media/photos')
from django_fixmystreet.fixmystreet.tests.views import *
from django_fixmystreet.fixmystreet.tests.reports import *
from django_fixmystreet.fixmystreet.tests.users import *
from django_fixmystreet.fixmystreet.tests.organisation_entity import *
from django_fixmystreet.fixmystreet.tests.mail import *
# from django_fixmystreet.fixmystreet.tests.api import *
## Instruction:
Fix unit test fixtures files
## Code After:
import shutil
import os
from django.core.files.storage import default_storage
from django.test import TestCase
class SampleFilesTestCase(TestCase):
fixtures = ['sample']
@classmethod
def setUpClass(cls):
default_storage.location = 'media' # force using source media folder to avoid real data erasing
# @classmethod
# def setUpClass(cls):
# shutil.copytree('media', 'media-tmp')
# default_storage.location = 'media-tmp'
#
# @classmethod
# def tearDownClass(self):
# shutil.rmtree('media-tmp')
def _fixture_setup(self):
if os.path.exists('media/photos'):
shutil.rmtree('media/photos')
shutil.copytree('media/photos-sample', 'media/photos')
super(SampleFilesTestCase, self)._fixture_setup()
from django_fixmystreet.fixmystreet.tests.views import *
from django_fixmystreet.fixmystreet.tests.reports import *
from django_fixmystreet.fixmystreet.tests.users import *
from django_fixmystreet.fixmystreet.tests.organisation_entity import *
from django_fixmystreet.fixmystreet.tests.mail import *
# from django_fixmystreet.fixmystreet.tests.api import *
| // ... existing code ...
super(SampleFilesTestCase, self)._fixture_setup()
from django_fixmystreet.fixmystreet.tests.views import *
// ... rest of the code ... |
bc8e548e51fddc251eb2e915883e3ee57bb9515b | zc_common/jwt_auth/utils.py | zc_common/jwt_auth/utils.py | import jwt
from rest_framework_jwt.settings import api_settings
def jwt_payload_handler(user):
# The handler from rest_framework_jwt removed user_id, so this is a fork
payload = {
'id': user.pk,
'roles': user.get_roles(),
}
return payload
def jwt_encode_handler(payload):
return jwt.encode(
payload,
api_settings.JWT_SECRET_KEY,
api_settings.JWT_ALGORITHM
).decode('utf-8')
| import jwt
from rest_framework_jwt.settings import api_settings
def jwt_payload_handler(user):
'''Constructs a payload for a user JWT. This is a slimmed down version of
https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11
:param User: an object with `pk` and `get_roles()`
:return: A dictionary that can be passed into `jwt_encode_handler`
'''
payload = {
'id': user.pk,
'roles': user.get_roles(),
}
return payload
def jwt_encode_handler(payload):
'''
Encodes a payload into a valid JWT.
:param payload: a dictionary
:return: an encoded JWT string
'''
return jwt.encode(
payload,
api_settings.JWT_SECRET_KEY,
api_settings.JWT_ALGORITHM
).decode('utf-8')
| Add docstrings to jwt handlers | Add docstrings to jwt handlers | Python | mit | ZeroCater/zc_common,ZeroCater/zc_common | import jwt
from rest_framework_jwt.settings import api_settings
def jwt_payload_handler(user):
- # The handler from rest_framework_jwt removed user_id, so this is a fork
+ '''Constructs a payload for a user JWT. This is a slimmed down version of
+ https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11
+
+ :param User: an object with `pk` and `get_roles()`
+ :return: A dictionary that can be passed into `jwt_encode_handler`
+ '''
+
payload = {
'id': user.pk,
'roles': user.get_roles(),
}
return payload
def jwt_encode_handler(payload):
+ '''
+ Encodes a payload into a valid JWT.
+
+ :param payload: a dictionary
+ :return: an encoded JWT string
+ '''
+
return jwt.encode(
payload,
api_settings.JWT_SECRET_KEY,
api_settings.JWT_ALGORITHM
).decode('utf-8')
| Add docstrings to jwt handlers | ## Code Before:
import jwt
from rest_framework_jwt.settings import api_settings
def jwt_payload_handler(user):
# The handler from rest_framework_jwt removed user_id, so this is a fork
payload = {
'id': user.pk,
'roles': user.get_roles(),
}
return payload
def jwt_encode_handler(payload):
return jwt.encode(
payload,
api_settings.JWT_SECRET_KEY,
api_settings.JWT_ALGORITHM
).decode('utf-8')
## Instruction:
Add docstrings to jwt handlers
## Code After:
import jwt
from rest_framework_jwt.settings import api_settings
def jwt_payload_handler(user):
'''Constructs a payload for a user JWT. This is a slimmed down version of
https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11
:param User: an object with `pk` and `get_roles()`
:return: A dictionary that can be passed into `jwt_encode_handler`
'''
payload = {
'id': user.pk,
'roles': user.get_roles(),
}
return payload
def jwt_encode_handler(payload):
'''
Encodes a payload into a valid JWT.
:param payload: a dictionary
:return: an encoded JWT string
'''
return jwt.encode(
payload,
api_settings.JWT_SECRET_KEY,
api_settings.JWT_ALGORITHM
).decode('utf-8')
| # ... existing code ...
def jwt_payload_handler(user):
'''Constructs a payload for a user JWT. This is a slimmed down version of
https://github.com/GetBlimp/django-rest-framework-jwt/blob/master/rest_framework_jwt/utils.py#L11
:param User: an object with `pk` and `get_roles()`
:return: A dictionary that can be passed into `jwt_encode_handler`
'''
payload = {
'id': user.pk,
# ... modified code ...
def jwt_encode_handler(payload):
'''
Encodes a payload into a valid JWT.
:param payload: a dictionary
:return: an encoded JWT string
'''
return jwt.encode(
payload,
# ... rest of the code ... |
0eca2340a4d38b542659505ba386a23129f8ac0b | google/cloud/speech/__init__.py | google/cloud/speech/__init__.py |
"""Google Cloud Speech API wrapper."""
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
|
"""Google Cloud Speech API wrapper."""
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
from google.cloud.speech.encoding import Encoding
| Make Encoding accessible from speech.Encoding. | Make Encoding accessible from speech.Encoding.
| Python | apache-2.0 | googleapis/python-speech,googleapis/python-speech |
"""Google Cloud Speech API wrapper."""
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
+ from google.cloud.speech.encoding import Encoding
| Make Encoding accessible from speech.Encoding. | ## Code Before:
"""Google Cloud Speech API wrapper."""
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
## Instruction:
Make Encoding accessible from speech.Encoding.
## Code After:
"""Google Cloud Speech API wrapper."""
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
from google.cloud.speech.encoding import Encoding
| // ... existing code ...
from google.cloud.speech.client import Client
from google.cloud.speech.connection import Connection
from google.cloud.speech.encoding import Encoding
// ... rest of the code ... |
4b84cedd15a2774391544a6edee3532e5e267608 | tests/docs/test_docs.py | tests/docs/test_docs.py | import subprocess
import unittest
import os
import subprocess
import unittest
import os
class Doc_Test(unittest.TestCase):
@property
def path_to_docs(self):
dirname, filename = os.path.split(os.path.abspath(__file__))
return dirname.split(os.path.sep)[:-2] + ['docs']
def test_html(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
os.chdir(wd)
def test_linkcheck(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "linkcheck"])
print(response.returncode)
self.assertTrue(response.returncode == 0)
os.chdir(wd)
if __name__ == '__main__':
unittest.main()
| import subprocess
import unittest
import os
import subprocess
import unittest
import os
class Doc_Test(unittest.TestCase):
@property
def path_to_docs(self):
dirname, filename = os.path.split(os.path.abspath(__file__))
return dirname.split(os.path.sep)[:-2] + ['docs']
def test_html(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows
# self.assertTrue(response == 0)
os.chdir(wd)
def test_linkcheck(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "linkcheck"])
print(response.returncode)
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows
# print(response)
# self.assertTrue(response == 0)
os.chdir(wd)
if __name__ == '__main__':
unittest.main()
| Edit docs test for local test on windows machine | Edit docs test for local test on windows machine
| Python | mit | simpeg/simpeg | import subprocess
import unittest
import os
import subprocess
import unittest
import os
class Doc_Test(unittest.TestCase):
@property
def path_to_docs(self):
dirname, filename = os.path.split(os.path.abspath(__file__))
return dirname.split(os.path.sep)[:-2] + ['docs']
def test_html(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
+ # response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows
+ # self.assertTrue(response == 0)
+
os.chdir(wd)
def test_linkcheck(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "linkcheck"])
print(response.returncode)
self.assertTrue(response.returncode == 0)
+ # response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows
+ # print(response)
+ # self.assertTrue(response == 0)
+
os.chdir(wd)
if __name__ == '__main__':
unittest.main()
| Edit docs test for local test on windows machine | ## Code Before:
import subprocess
import unittest
import os
import subprocess
import unittest
import os
class Doc_Test(unittest.TestCase):
@property
def path_to_docs(self):
dirname, filename = os.path.split(os.path.abspath(__file__))
return dirname.split(os.path.sep)[:-2] + ['docs']
def test_html(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
os.chdir(wd)
def test_linkcheck(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "linkcheck"])
print(response.returncode)
self.assertTrue(response.returncode == 0)
os.chdir(wd)
if __name__ == '__main__':
unittest.main()
## Instruction:
Edit docs test for local test on windows machine
## Code After:
import subprocess
import unittest
import os
import subprocess
import unittest
import os
class Doc_Test(unittest.TestCase):
@property
def path_to_docs(self):
dirname, filename = os.path.split(os.path.abspath(__file__))
return dirname.split(os.path.sep)[:-2] + ['docs']
def test_html(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows
# self.assertTrue(response == 0)
os.chdir(wd)
def test_linkcheck(self):
wd = os.getcwd()
os.chdir(os.path.sep.join(self.path_to_docs))
response = subprocess.run(["make", "linkcheck"])
print(response.returncode)
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows
# print(response)
# self.assertTrue(response == 0)
os.chdir(wd)
if __name__ == '__main__':
unittest.main()
| ...
response = subprocess.run(["make", "html"])
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "html"], shell=True) # Needed for local test on Windows
# self.assertTrue(response == 0)
os.chdir(wd)
...
print(response.returncode)
self.assertTrue(response.returncode == 0)
# response = subprocess.call(["make", "linkcheck"], shell=True) # Needed for local test on Windows
# print(response)
# self.assertTrue(response == 0)
os.chdir(wd)
... |
7d3de3aa2441739aa951aa100c057cfa878887d5 | nukedb.py | nukedb.py | import sqlite3
if __name__=="__main__":
conn = sqlite3.connect('auxgis.db')
c = conn.cursor()
try:
c.execute('''DROP TABLE pos;''')
except:
pass
try:
c.execute('''DROP TABLE data;''')
except:
pass
conn.commit()
| import sqlite3
if __name__=="__main__":
conn = sqlite3.connect('auxgis.db')
c = conn.cursor()
try:
c.execute('''DROP TABLE pos;''')
except:
pass
try:
c.execute('''DROP TABLE data;''')
except:
pass
try:
c.execute('''DROP TABLE recentchanges;''')
except:
pass
conn.commit()
| Drop recent changes on nuke | Drop recent changes on nuke
| Python | bsd-3-clause | TimSC/auxgis | import sqlite3
if __name__=="__main__":
conn = sqlite3.connect('auxgis.db')
c = conn.cursor()
try:
c.execute('''DROP TABLE pos;''')
except:
pass
try:
c.execute('''DROP TABLE data;''')
except:
pass
+ try:
+ c.execute('''DROP TABLE recentchanges;''')
+ except:
+ pass
conn.commit()
| Drop recent changes on nuke | ## Code Before:
import sqlite3
if __name__=="__main__":
conn = sqlite3.connect('auxgis.db')
c = conn.cursor()
try:
c.execute('''DROP TABLE pos;''')
except:
pass
try:
c.execute('''DROP TABLE data;''')
except:
pass
conn.commit()
## Instruction:
Drop recent changes on nuke
## Code After:
import sqlite3
if __name__=="__main__":
conn = sqlite3.connect('auxgis.db')
c = conn.cursor()
try:
c.execute('''DROP TABLE pos;''')
except:
pass
try:
c.execute('''DROP TABLE data;''')
except:
pass
try:
c.execute('''DROP TABLE recentchanges;''')
except:
pass
conn.commit()
| // ... existing code ...
except:
pass
try:
c.execute('''DROP TABLE recentchanges;''')
except:
pass
conn.commit()
// ... rest of the code ... |
4922d53f95b3f7c055afe1d0af0088b505cbc0d2 | addons/bestja_configuration_ucw/__openerp__.py | addons/bestja_configuration_ucw/__openerp__.py | {
'name': "Bestja: UCW",
'summary': "Installation configuration for UCW",
'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu",
'author': "Laboratorium EE",
'website': "http://www.laboratorium.ee",
'version': '0.1',
'category': 'Specific Industry Applications',
'depends': [
'base',
'bestja_base',
'bestja_volunteer',
'bestja_volunteer_notes',
'bestja_account_deletion',
'bestja_organization',
'bestja_project',
'bestja_offers',
'bestja_offers_moderation',
'bestja_offers_invitations',
'bestja_offers_categorization',
'bestja_files',
'bestja_application_moderation',
'bestja_ucw_permissions',
],
'data': [
'data.xml',
],
'application': True,
}
| {
'name': "Bestja: UCW",
'summary': "Installation configuration for UCW",
'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu",
'author': "Laboratorium EE",
'website': "http://www.laboratorium.ee",
'version': '0.1',
'category': 'Specific Industry Applications',
'depends': [
'base',
'website_blog',
'bestja_base',
'bestja_volunteer',
'bestja_volunteer_notes',
'bestja_account_deletion',
'bestja_organization',
'bestja_project',
'bestja_offers',
'bestja_offers_moderation',
'bestja_offers_invitations',
'bestja_offers_categorization',
'bestja_files',
'bestja_application_moderation',
'bestja_ucw_permissions',
],
'data': [
'data.xml',
],
'application': True,
}
| Enable Odoo blog for UCW | Enable Odoo blog for UCW
| Python | agpl-3.0 | EE/bestja,EE/bestja,KamilWo/bestja,KamilWo/bestja,KrzysiekJ/bestja,ludwiktrammer/bestja,EE/bestja,ludwiktrammer/bestja,KamilWo/bestja,KrzysiekJ/bestja,ludwiktrammer/bestja,KrzysiekJ/bestja | {
'name': "Bestja: UCW",
'summary': "Installation configuration for UCW",
'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu",
'author': "Laboratorium EE",
'website': "http://www.laboratorium.ee",
'version': '0.1',
'category': 'Specific Industry Applications',
'depends': [
'base',
+ 'website_blog',
'bestja_base',
'bestja_volunteer',
'bestja_volunteer_notes',
'bestja_account_deletion',
'bestja_organization',
'bestja_project',
'bestja_offers',
'bestja_offers_moderation',
'bestja_offers_invitations',
'bestja_offers_categorization',
'bestja_files',
'bestja_application_moderation',
'bestja_ucw_permissions',
],
'data': [
'data.xml',
],
'application': True,
}
| Enable Odoo blog for UCW | ## Code Before:
{
'name': "Bestja: UCW",
'summary': "Installation configuration for UCW",
'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu",
'author': "Laboratorium EE",
'website': "http://www.laboratorium.ee",
'version': '0.1',
'category': 'Specific Industry Applications',
'depends': [
'base',
'bestja_base',
'bestja_volunteer',
'bestja_volunteer_notes',
'bestja_account_deletion',
'bestja_organization',
'bestja_project',
'bestja_offers',
'bestja_offers_moderation',
'bestja_offers_invitations',
'bestja_offers_categorization',
'bestja_files',
'bestja_application_moderation',
'bestja_ucw_permissions',
],
'data': [
'data.xml',
],
'application': True,
}
## Instruction:
Enable Odoo blog for UCW
## Code After:
{
'name': "Bestja: UCW",
'summary': "Installation configuration for UCW",
'description': "Installation configuration for Uniwersyteckie Centrum Wolontariatu",
'author': "Laboratorium EE",
'website': "http://www.laboratorium.ee",
'version': '0.1',
'category': 'Specific Industry Applications',
'depends': [
'base',
'website_blog',
'bestja_base',
'bestja_volunteer',
'bestja_volunteer_notes',
'bestja_account_deletion',
'bestja_organization',
'bestja_project',
'bestja_offers',
'bestja_offers_moderation',
'bestja_offers_invitations',
'bestja_offers_categorization',
'bestja_files',
'bestja_application_moderation',
'bestja_ucw_permissions',
],
'data': [
'data.xml',
],
'application': True,
}
| // ... existing code ...
'depends': [
'base',
'website_blog',
'bestja_base',
'bestja_volunteer',
// ... rest of the code ... |
8a7837a8ce7b35c3141374c6a5c99361261fa70a | Cura/avr_isp/chipDB.py | Cura/avr_isp/chipDB.py |
avrChipDB = {
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
'pageSize': 128,
'pageCount': 1024,
},
}
def getChipFromDB(sig):
for chip in avrChipDB.values():
if chip['signature'] == sig:
return chip
return False
|
avrChipDB = {
'ATMega1280': {
'signature': [0x1E, 0x97, 0x03],
'pageSize': 128,
'pageCount': 512,
},
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
'pageSize': 128,
'pageCount': 1024,
},
}
def getChipFromDB(sig):
for chip in avrChipDB.values():
if chip['signature'] == sig:
return chip
return False
| Add ATMega1280 chip to programmer chips. | Add ATMega1280 chip to programmer chips.
| Python | agpl-3.0 | MolarAmbiguity/OctoPrint,EZ3-India/EZ-Remote,JackGavin13/octoprint-test-not-finished,spapadim/OctoPrint,dragondgold/OctoPrint,hudbrog/OctoPrint,CapnBry/OctoPrint,Javierma/OctoPrint-TFG,chriskoz/OctoPrint,javivi001/OctoPrint,shohei/Octoprint,eddieparker/OctoPrint,MolarAmbiguity/OctoPrint,mayoff/OctoPrint,uuv/OctoPrint,C-o-r-E/OctoPrint,Mikk36/OctoPrint,DanLipsitt/OctoPrint,shohei/Octoprint,beeverycreative/BEEweb,alex1818/OctoPrint,EZ3-India/EZ-Remote,alex1818/OctoPrint,shohei/Octoprint,markwal/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,nicanor-romero/OctoPrint,punkkeks/OctoPrint,d42/octoprint-fork,Javierma/OctoPrint-TFG,3dprintcanalhouse/octoprint2,ErikDeBruijn/OctoPrint,punkkeks/OctoPrint,masterhou/OctoPrint,shaggythesheep/OctoPrint,chriskoz/OctoPrint,madhuni/AstroBox,Catrodigious/OctoPrint-TAM,alephobjects/Cura,javivi001/OctoPrint,uuv/OctoPrint,leductan-nguyen/RaionPi,MoonshineSG/OctoPrint,eliasbakken/OctoPrint,nicanor-romero/OctoPrint,Skeen/OctoPrint,javivi001/OctoPrint,Salandora/OctoPrint,jneves/OctoPrint,hudbrog/OctoPrint,shaggythesheep/OctoPrint,MoonshineSG/OctoPrint,skieast/OctoPrint,abinashk-inf/AstroBox,nickverschoor/OctoPrint,eddieparker/OctoPrint,EZ3-India/EZ-Remote,EZ3-India/EZ-Remote,abinashk-inf/AstroBox,mrbeam/OctoPrint,abinashk-inf/AstroBox,mrbeam/OctoPrint,Voxel8/OctoPrint,sstocker46/OctoPrint,bicephale/OctoPrint,dragondgold/OctoPrint,Jaesin/OctoPrint,mcanes/OctoPrint,ryanneufeld/OctoPrint,Salandora/OctoPrint,CapnBry/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,alephobjects/Cura,mcanes/OctoPrint,markwal/OctoPrint,sstocker46/OctoPrint,Jaesin/OctoPrint,3dprintcanalhouse/octoprint1,skieast/OctoPrint,madhuni/AstroBox,markwal/OctoPrint,Mikk36/OctoPrint,AstroPrint/AstroBox,ymilord/OctoPrint-MrBeam,dansantee/OctoPrint,Jaesin/OctoPrint,punkkeks/OctoPrint,ymilord/OctoPrint-MrBeam,rurkowce/octoprint-fork,foosel/OctoPrint,Salandora/OctoPrint,spapadim/OctoPrint,MoonshineSG/OctoPrint,spapadim/OctoPrint,madhuni/AstroBox,masterhou/OctoPrint,ymilord/OctoPrint-MrBeam,alephobjects/Cura,ryanneufeld/OctoPrint,chriskoz/OctoPrint,hudbrog/OctoPrint,Mikk36/OctoPrint,eddieparker/OctoPrint,leductan-nguyen/RaionPi,JackGavin13/octoprint-test-not-finished,beeverycreative/BEEweb,bicephale/OctoPrint,nicanor-romero/OctoPrint,jneves/OctoPrint,JackGavin13/octoprint-test-not-finished,ErikDeBruijn/OctoPrint,leductan-nguyen/RaionPi,CapnBry/OctoPrint,chriskoz/OctoPrint,ryanneufeld/OctoPrint,3dprintcanalhouse/octoprint1,mrbeam/OctoPrint,senttech/OctoPrint,Javierma/OctoPrint-TFG,dansantee/OctoPrint,Voxel8/OctoPrint,bicephale/OctoPrint,MolarAmbiguity/OctoPrint,MaxOLydian/OctoPrint,eliasbakken/OctoPrint,DanLipsitt/OctoPrint,mayoff/OctoPrint,Skeen/OctoPrint,Jaesin/OctoPrint,rurkowce/octoprint-fork,CapnBry/OctoPrint,AstroPrint/AstroBox,madhuni/AstroBox,uuv/OctoPrint,abinashk-inf/AstroBox,JackGavin13/octoprint-test-not-finished,SeveQ/OctoPrint,sstocker46/OctoPrint,dansantee/OctoPrint,skieast/OctoPrint,mayoff/OctoPrint,C-o-r-E/OctoPrint,eliasbakken/OctoPrint,ryanneufeld/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,bicephale/OctoPrint,SeveQ/OctoPrint,MoonshineSG/OctoPrint,SeveQ/OctoPrint,senttech/OctoPrint,shohei/Octoprint,ymilord/OctoPrint-MrBeam,3dprintcanalhouse/octoprint2,d42/octoprint-fork,mcanes/OctoPrint,Voxel8/OctoPrint,senttech/OctoPrint,ymilord/OctoPrint-MrBeam,leductan-nguyen/RaionPi,Javierma/OctoPrint-TFG,Salandora/OctoPrint,C-o-r-E/OctoPrint,alex1818/OctoPrint,MaxOLydian/OctoPrint,shaggythesheep/OctoPrint,masterhou/OctoPrint,shohei/Octoprint,ErikDeBruijn/OctoPrint,jneves/OctoPrint,Catrodigious/OctoPrint-TAM,foosel/OctoPrint,dragondgold/OctoPrint,senttech/OctoPrint,aerickson/OctoPrint,MaxOLydian/OctoPrint,nickverschoor/OctoPrint,Skeen/OctoPrint,Catrodigious/OctoPrint-TAM,AstroPrint/AstroBox |
avrChipDB = {
+ 'ATMega1280': {
+ 'signature': [0x1E, 0x97, 0x03],
+ 'pageSize': 128,
+ 'pageCount': 512,
+ },
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
'pageSize': 128,
'pageCount': 1024,
},
}
def getChipFromDB(sig):
for chip in avrChipDB.values():
if chip['signature'] == sig:
return chip
return False
| Add ATMega1280 chip to programmer chips. | ## Code Before:
avrChipDB = {
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
'pageSize': 128,
'pageCount': 1024,
},
}
def getChipFromDB(sig):
for chip in avrChipDB.values():
if chip['signature'] == sig:
return chip
return False
## Instruction:
Add ATMega1280 chip to programmer chips.
## Code After:
avrChipDB = {
'ATMega1280': {
'signature': [0x1E, 0x97, 0x03],
'pageSize': 128,
'pageCount': 512,
},
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
'pageSize': 128,
'pageCount': 1024,
},
}
def getChipFromDB(sig):
for chip in avrChipDB.values():
if chip['signature'] == sig:
return chip
return False
| // ... existing code ...
avrChipDB = {
'ATMega1280': {
'signature': [0x1E, 0x97, 0x03],
'pageSize': 128,
'pageCount': 512,
},
'ATMega2560': {
'signature': [0x1E, 0x98, 0x01],
// ... rest of the code ... |
874a6eff186d1c1ca6f90d69fd24fad11180c5a9 | thread_output_ctrl.py | thread_output_ctrl.py | import threading
import wx
from styled_text_ctrl import StyledTextCtrl
class ThreadOutputCtrl(StyledTextCtrl):
def __init__(self, parent, env, auto_scroll=False):
StyledTextCtrl.__init__(self, parent, env)
self.auto_scroll = auto_scroll
self.__lock = threading.Lock()
self.__queue = []
self.__timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer)
def __OnTimer(self, evt):
self.flush()
def flush(self):
with self.__lock:
queue, self.__queue = self.__queue, []
lines = "".join(queue)
if lines:
with self.ModifyReadOnly():
self.AppendText(lines)
self.EmptyUndoBuffer()
if self.auto_scroll:
self.ScrollToLine(self.GetLineCount() - 1)
def start(self, interval=100):
self.SetReadOnly(True)
self.__timer.Start(interval)
def stop(self):
self.__timer.Stop()
self.flush()
self.SetReadOnly(False)
def write(self, s):
with self.__lock:
self.__queue.append(s)
def ClearAll(self):
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
| import threading
import wx
from styled_text_ctrl import StyledTextCtrl
class ThreadOutputCtrl(StyledTextCtrl):
def __init__(self, parent, env, auto_scroll=False):
StyledTextCtrl.__init__(self, parent, env)
self.auto_scroll = auto_scroll
self.__lock = threading.Lock()
self.__queue = []
self.__timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer)
def __OnTimer(self, evt):
self.flush()
def flush(self):
with self.__lock:
queue, self.__queue = self.__queue, []
lines = "".join(queue)
if lines:
with self.ModifyReadOnly():
self.AppendText(lines)
self.EmptyUndoBuffer()
if self.auto_scroll:
self.ScrollToLine(self.GetLineCount() - 1)
def start(self, interval=100):
self.SetReadOnly(True)
self.__timer.Start(interval)
def stop(self):
self.__timer.Stop()
self.flush()
self.SetReadOnly(False)
def write(self, s):
with self.__lock:
self.__queue.append(s)
def ClearAll(self):
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
self.EmptyUndoBuffer()
| Clear undo buffer when terminal cleared. | Clear undo buffer when terminal cleared.
| Python | mit | shaurz/devo | import threading
import wx
from styled_text_ctrl import StyledTextCtrl
class ThreadOutputCtrl(StyledTextCtrl):
def __init__(self, parent, env, auto_scroll=False):
StyledTextCtrl.__init__(self, parent, env)
self.auto_scroll = auto_scroll
self.__lock = threading.Lock()
self.__queue = []
self.__timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer)
def __OnTimer(self, evt):
self.flush()
def flush(self):
with self.__lock:
queue, self.__queue = self.__queue, []
lines = "".join(queue)
if lines:
with self.ModifyReadOnly():
self.AppendText(lines)
self.EmptyUndoBuffer()
if self.auto_scroll:
self.ScrollToLine(self.GetLineCount() - 1)
def start(self, interval=100):
self.SetReadOnly(True)
self.__timer.Start(interval)
def stop(self):
self.__timer.Stop()
self.flush()
self.SetReadOnly(False)
def write(self, s):
with self.__lock:
self.__queue.append(s)
def ClearAll(self):
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
+ self.EmptyUndoBuffer()
| Clear undo buffer when terminal cleared. | ## Code Before:
import threading
import wx
from styled_text_ctrl import StyledTextCtrl
class ThreadOutputCtrl(StyledTextCtrl):
def __init__(self, parent, env, auto_scroll=False):
StyledTextCtrl.__init__(self, parent, env)
self.auto_scroll = auto_scroll
self.__lock = threading.Lock()
self.__queue = []
self.__timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer)
def __OnTimer(self, evt):
self.flush()
def flush(self):
with self.__lock:
queue, self.__queue = self.__queue, []
lines = "".join(queue)
if lines:
with self.ModifyReadOnly():
self.AppendText(lines)
self.EmptyUndoBuffer()
if self.auto_scroll:
self.ScrollToLine(self.GetLineCount() - 1)
def start(self, interval=100):
self.SetReadOnly(True)
self.__timer.Start(interval)
def stop(self):
self.__timer.Stop()
self.flush()
self.SetReadOnly(False)
def write(self, s):
with self.__lock:
self.__queue.append(s)
def ClearAll(self):
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
## Instruction:
Clear undo buffer when terminal cleared.
## Code After:
import threading
import wx
from styled_text_ctrl import StyledTextCtrl
class ThreadOutputCtrl(StyledTextCtrl):
def __init__(self, parent, env, auto_scroll=False):
StyledTextCtrl.__init__(self, parent, env)
self.auto_scroll = auto_scroll
self.__lock = threading.Lock()
self.__queue = []
self.__timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.__OnTimer, self.__timer)
def __OnTimer(self, evt):
self.flush()
def flush(self):
with self.__lock:
queue, self.__queue = self.__queue, []
lines = "".join(queue)
if lines:
with self.ModifyReadOnly():
self.AppendText(lines)
self.EmptyUndoBuffer()
if self.auto_scroll:
self.ScrollToLine(self.GetLineCount() - 1)
def start(self, interval=100):
self.SetReadOnly(True)
self.__timer.Start(interval)
def stop(self):
self.__timer.Stop()
self.flush()
self.SetReadOnly(False)
def write(self, s):
with self.__lock:
self.__queue.append(s)
def ClearAll(self):
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
self.EmptyUndoBuffer()
| ...
with self.ModifyReadOnly():
StyledTextCtrl.ClearAll(self)
self.EmptyUndoBuffer()
... |
4d7c1fec37943558ccc8bf6a17860b2a86fe1941 | gee_asset_manager/batch_copy.py | gee_asset_manager/batch_copy.py | import ee
import os
import csv
import logging
def copy(source, destination):
with open(source, 'r') as f:
reader = csv.reader(f)
for line in reader:
name = line[0]
gme_id = line[1]
gme_path = 'GME/images/' + gme_id
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
ee.data.copyAsset(gme_path, ee_path)
if __name__ == '__main__':
ee.Initialize()
assets = '/home/tracek/Data/consbio2016/test.csv'
with open(assets, 'r') as f:
reader = csv.reader(f) | import ee
import os
import csv
import logging
def copy(source, destination):
with open(source, 'r') as f:
reader = csv.reader(f)
for line in reader:
name = line[0]
gme_id = line[1]
gme_path = 'GME/images/' + gme_id
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
try:
ee.data.copyAsset(gme_path, ee_path)
except ee.EEException as e:
with open('failed_batch_copy.csv', 'w') as fout:
fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e)
if __name__ == '__main__':
ee.Initialize()
assets = '/home/tracek/Data/consbio2016/test.csv'
with open(assets, 'r') as f:
reader = csv.reader(f) | Add exception handling to batch copy | Add exception handling to batch copy
| Python | apache-2.0 | tracek/gee_asset_manager | import ee
import os
import csv
import logging
def copy(source, destination):
with open(source, 'r') as f:
reader = csv.reader(f)
for line in reader:
name = line[0]
gme_id = line[1]
gme_path = 'GME/images/' + gme_id
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
+ try:
- ee.data.copyAsset(gme_path, ee_path)
+ ee.data.copyAsset(gme_path, ee_path)
+ except ee.EEException as e:
+ with open('failed_batch_copy.csv', 'w') as fout:
+ fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e)
if __name__ == '__main__':
ee.Initialize()
assets = '/home/tracek/Data/consbio2016/test.csv'
with open(assets, 'r') as f:
reader = csv.reader(f) | Add exception handling to batch copy | ## Code Before:
import ee
import os
import csv
import logging
def copy(source, destination):
with open(source, 'r') as f:
reader = csv.reader(f)
for line in reader:
name = line[0]
gme_id = line[1]
gme_path = 'GME/images/' + gme_id
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
ee.data.copyAsset(gme_path, ee_path)
if __name__ == '__main__':
ee.Initialize()
assets = '/home/tracek/Data/consbio2016/test.csv'
with open(assets, 'r') as f:
reader = csv.reader(f)
## Instruction:
Add exception handling to batch copy
## Code After:
import ee
import os
import csv
import logging
def copy(source, destination):
with open(source, 'r') as f:
reader = csv.reader(f)
for line in reader:
name = line[0]
gme_id = line[1]
gme_path = 'GME/images/' + gme_id
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
try:
ee.data.copyAsset(gme_path, ee_path)
except ee.EEException as e:
with open('failed_batch_copy.csv', 'w') as fout:
fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e)
if __name__ == '__main__':
ee.Initialize()
assets = '/home/tracek/Data/consbio2016/test.csv'
with open(assets, 'r') as f:
reader = csv.reader(f) | # ... existing code ...
ee_path = os.path.join(destination, name)
logging.info('Copying asset %s to %s', gme_path, ee_path)
try:
ee.data.copyAsset(gme_path, ee_path)
except ee.EEException as e:
with open('failed_batch_copy.csv', 'w') as fout:
fout.write('%s,%s,%s,%s', name, gme_id, ee_path,e)
# ... rest of the code ... |
9b54d728a245855cba724a91d372a15a4f4abb6d | shop/checkout/models.py | shop/checkout/models.py | """Checkout Models"""
import functools
from flask import redirect, url_for
from fulfil_client.model import ModelType, StringType
from shop.fulfilio import Model
from shop.globals import current_cart, current_channel
def not_empty_cart(function):
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.is_empty:
return redirect(url_for('cart.view_cart'))
return function(*args, **kwargs)
return wrapper
def sale_has_non_guest_party(function):
"""
Ensure that the sale has a party who is not guest.
The sign-in method authomatically changes the party to a party based on the
session.
"""
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.sale and cart.sale.party and \
cart.sale.party.id == current_channel.anonymous_customer.id:
return redirect(url_for('checkout.sign_in'))
return function(*args, **kwargs)
return wrapper
class PaymentGateway(Model):
__model_name__ = 'payment_gateway.gateway'
provider = StringType()
stripe_publishable_key = StringType()
class PaymentProfile(Model):
__model_name__ = 'party.payment_profile'
party = ModelType('party.party')
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
rec_name = StringType()
| """Checkout Models"""
import functools
from flask import redirect, url_for
from fulfil_client.model import ModelType, StringType
from shop.fulfilio import Model
from shop.globals import current_cart, current_channel
def not_empty_cart(function):
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.is_empty:
return redirect(url_for('cart.view_cart'))
return function(*args, **kwargs)
return wrapper
def sale_has_non_guest_party(function):
"""
Ensure that the sale has a party who is not guest.
The sign-in method authomatically changes the party to a party based on the
session.
"""
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.sale and cart.sale.party and \
cart.sale.party.id == current_channel.anonymous_customer.id:
return redirect(url_for('checkout.sign_in'))
return function(*args, **kwargs)
return wrapper
class PaymentGateway(Model):
__model_name__ = 'payment_gateway.gateway'
provider = StringType()
stripe_publishable_key = StringType()
class PaymentProfile(Model):
__model_name__ = 'party.payment_profile'
party = ModelType('party.party')
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
expiry_month = StringType()
expiry_year = StringType()
rec_name = StringType()
| Add expiry fields on card model | Add expiry fields on card model
| Python | bsd-3-clause | joeirimpan/shop,joeirimpan/shop,joeirimpan/shop | """Checkout Models"""
import functools
from flask import redirect, url_for
from fulfil_client.model import ModelType, StringType
from shop.fulfilio import Model
from shop.globals import current_cart, current_channel
def not_empty_cart(function):
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.is_empty:
return redirect(url_for('cart.view_cart'))
return function(*args, **kwargs)
return wrapper
def sale_has_non_guest_party(function):
"""
Ensure that the sale has a party who is not guest.
The sign-in method authomatically changes the party to a party based on the
session.
"""
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.sale and cart.sale.party and \
cart.sale.party.id == current_channel.anonymous_customer.id:
return redirect(url_for('checkout.sign_in'))
return function(*args, **kwargs)
return wrapper
class PaymentGateway(Model):
__model_name__ = 'payment_gateway.gateway'
provider = StringType()
stripe_publishable_key = StringType()
class PaymentProfile(Model):
__model_name__ = 'party.payment_profile'
party = ModelType('party.party')
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
+ expiry_month = StringType()
+ expiry_year = StringType()
rec_name = StringType()
| Add expiry fields on card model | ## Code Before:
"""Checkout Models"""
import functools
from flask import redirect, url_for
from fulfil_client.model import ModelType, StringType
from shop.fulfilio import Model
from shop.globals import current_cart, current_channel
def not_empty_cart(function):
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.is_empty:
return redirect(url_for('cart.view_cart'))
return function(*args, **kwargs)
return wrapper
def sale_has_non_guest_party(function):
"""
Ensure that the sale has a party who is not guest.
The sign-in method authomatically changes the party to a party based on the
session.
"""
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.sale and cart.sale.party and \
cart.sale.party.id == current_channel.anonymous_customer.id:
return redirect(url_for('checkout.sign_in'))
return function(*args, **kwargs)
return wrapper
class PaymentGateway(Model):
__model_name__ = 'payment_gateway.gateway'
provider = StringType()
stripe_publishable_key = StringType()
class PaymentProfile(Model):
__model_name__ = 'party.payment_profile'
party = ModelType('party.party')
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
rec_name = StringType()
## Instruction:
Add expiry fields on card model
## Code After:
"""Checkout Models"""
import functools
from flask import redirect, url_for
from fulfil_client.model import ModelType, StringType
from shop.fulfilio import Model
from shop.globals import current_cart, current_channel
def not_empty_cart(function):
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.is_empty:
return redirect(url_for('cart.view_cart'))
return function(*args, **kwargs)
return wrapper
def sale_has_non_guest_party(function):
"""
Ensure that the sale has a party who is not guest.
The sign-in method authomatically changes the party to a party based on the
session.
"""
@functools.wraps(function)
def wrapper(*args, **kwargs):
cart = current_cart
if cart.sale and cart.sale.party and \
cart.sale.party.id == current_channel.anonymous_customer.id:
return redirect(url_for('checkout.sign_in'))
return function(*args, **kwargs)
return wrapper
class PaymentGateway(Model):
__model_name__ = 'payment_gateway.gateway'
provider = StringType()
stripe_publishable_key = StringType()
class PaymentProfile(Model):
__model_name__ = 'party.payment_profile'
party = ModelType('party.party')
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
expiry_month = StringType()
expiry_year = StringType()
rec_name = StringType()
| // ... existing code ...
gateway = ModelType('payment_gateway.gateway')
last_4_digits = StringType()
expiry_month = StringType()
expiry_year = StringType()
rec_name = StringType()
// ... rest of the code ... |
8df3076b6315a74e57ee27fe3478d36737be0ff9 | roche/scripts/xml-load.py | roche/scripts/xml-load.py |
import sys
import os
sys.path.append('.')
import roche.settings
from os import walk
from eulexistdb.db import ExistDB
from roche.settings import EXISTDB_SERVER_URL
#
# Timeout higher?
#
xmldb = ExistDB(timeout=30)
xmldb.createCollection('docker', True)
xmldb.createCollection('docker/texts', True)
os.chdir('../dublin-store')
for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'):
xmldb.createCollection('docker/texts' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
|
import sys
import os
sys.path.append('.')
import roche.settings
from os import walk
from eulexistdb.db import ExistDB
from roche.settings import EXISTDB_SERVER_URL
#
# Timeout higher?
#
xmldb = ExistDB(timeout=30)
xmldb.createCollection('docker', True)
xmldb.createCollection('docker/texts', True)
os.chdir('../dublin-store')
for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'):
xmldb.createCollection('docker/texts' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
#
# Load resources
#
for (dirpath, dirnames, filenames) in walk('resources'):
xmldb.createCollection('docker' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
| Load other resources into exist-db | Load other resources into exist-db
| Python | mit | beijingren/roche-website,beijingren/roche-website,beijingren/roche-website,beijingren/roche-website |
import sys
import os
sys.path.append('.')
import roche.settings
from os import walk
from eulexistdb.db import ExistDB
from roche.settings import EXISTDB_SERVER_URL
#
# Timeout higher?
#
xmldb = ExistDB(timeout=30)
xmldb.createCollection('docker', True)
xmldb.createCollection('docker/texts', True)
os.chdir('../dublin-store')
for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'):
xmldb.createCollection('docker/texts' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
+ #
+ # Load resources
+ #
+ for (dirpath, dirnames, filenames) in walk('resources'):
+ xmldb.createCollection('docker' + '/' + dirpath, True)
+ if filenames:
+ for filename in filenames:
+ with open(dirpath + '/' + filename) as f:
+ xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
+ | Load other resources into exist-db | ## Code Before:
import sys
import os
sys.path.append('.')
import roche.settings
from os import walk
from eulexistdb.db import ExistDB
from roche.settings import EXISTDB_SERVER_URL
#
# Timeout higher?
#
xmldb = ExistDB(timeout=30)
xmldb.createCollection('docker', True)
xmldb.createCollection('docker/texts', True)
os.chdir('../dublin-store')
for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'):
xmldb.createCollection('docker/texts' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
## Instruction:
Load other resources into exist-db
## Code After:
import sys
import os
sys.path.append('.')
import roche.settings
from os import walk
from eulexistdb.db import ExistDB
from roche.settings import EXISTDB_SERVER_URL
#
# Timeout higher?
#
xmldb = ExistDB(timeout=30)
xmldb.createCollection('docker', True)
xmldb.createCollection('docker/texts', True)
os.chdir('../dublin-store')
for (dirpath, dirnames, filenames) in walk('浙江大學圖書館'):
xmldb.createCollection('docker/texts' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
#
# Load resources
#
for (dirpath, dirnames, filenames) in walk('resources'):
xmldb.createCollection('docker' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
| # ... existing code ...
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker/texts' + '/' + dirpath + '/' + filename, True)
#
# Load resources
#
for (dirpath, dirnames, filenames) in walk('resources'):
xmldb.createCollection('docker' + '/' + dirpath, True)
if filenames:
for filename in filenames:
with open(dirpath + '/' + filename) as f:
xmldb.load(f, 'docker' + '/' + dirpath + '/' + filename, True)
# ... rest of the code ... |
61ca14440f39106b6109b96919b520e40170b1f3 | examples/tour_examples/xkcd_tour.py | examples/tour_examples/xkcd_tour.py | from seleniumbase import BaseCase
class MyTestClass(BaseCase):
def test_basic(self):
self.open('https://xkcd.com/1117/')
self.assert_element('img[alt="My Sky"]')
self.create_shepherd_tour()
self.add_tour_step("Welcome to XKCD!")
self.add_tour_step("This is the XKCD logo.", "#masthead img")
self.add_tour_step("Here's the daily webcomic.", "#comic img")
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
self.add_tour_step("Click for the license here.", 'a[rel="license"]')
self.add_tour_step("This selects a random comic.", 'a[href*="random"]')
self.add_tour_step("Thanks for taking this tour!")
# self.export_tour() # Use this to export the tour as [my_tour.js]
self.export_tour(filename="xkcd_tour.js") # You can customize the name
self.play_tour()
| from seleniumbase import BaseCase
class MyTestClass(BaseCase):
def test_basic(self):
self.open('https://xkcd.com/1117/')
self.assert_element('img[alt="My Sky"]')
self.create_shepherd_tour()
self.add_tour_step("Welcome to XKCD!")
self.add_tour_step("This is the XKCD logo.", "#masthead img")
self.add_tour_step("Here's the daily webcomic.", "#comic img")
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
self.add_tour_step("Click here for the license.", 'a[rel="license"]')
self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]')
self.add_tour_step("Thanks for taking this tour!")
self.export_tour(filename="xkcd_tour.js") # This exports the tour
self.play_tour() # This plays the tour
| Update a SeleniumBase tour example | Update a SeleniumBase tour example
| Python | mit | mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase | from seleniumbase import BaseCase
class MyTestClass(BaseCase):
def test_basic(self):
self.open('https://xkcd.com/1117/')
self.assert_element('img[alt="My Sky"]')
self.create_shepherd_tour()
self.add_tour_step("Welcome to XKCD!")
self.add_tour_step("This is the XKCD logo.", "#masthead img")
self.add_tour_step("Here's the daily webcomic.", "#comic img")
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
- self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]')
+ self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
- self.add_tour_step("Click for the license here.", 'a[rel="license"]')
+ self.add_tour_step("Click here for the license.", 'a[rel="license"]')
- self.add_tour_step("This selects a random comic.", 'a[href*="random"]')
+ self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]')
self.add_tour_step("Thanks for taking this tour!")
- # self.export_tour() # Use this to export the tour as [my_tour.js]
- self.export_tour(filename="xkcd_tour.js") # You can customize the name
+ self.export_tour(filename="xkcd_tour.js") # This exports the tour
- self.play_tour()
+ self.play_tour() # This plays the tour
| Update a SeleniumBase tour example | ## Code Before:
from seleniumbase import BaseCase
class MyTestClass(BaseCase):
def test_basic(self):
self.open('https://xkcd.com/1117/')
self.assert_element('img[alt="My Sky"]')
self.create_shepherd_tour()
self.add_tour_step("Welcome to XKCD!")
self.add_tour_step("This is the XKCD logo.", "#masthead img")
self.add_tour_step("Here's the daily webcomic.", "#comic img")
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
self.add_tour_step("Or here for the previous comic.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
self.add_tour_step("Click for the license here.", 'a[rel="license"]')
self.add_tour_step("This selects a random comic.", 'a[href*="random"]')
self.add_tour_step("Thanks for taking this tour!")
# self.export_tour() # Use this to export the tour as [my_tour.js]
self.export_tour(filename="xkcd_tour.js") # You can customize the name
self.play_tour()
## Instruction:
Update a SeleniumBase tour example
## Code After:
from seleniumbase import BaseCase
class MyTestClass(BaseCase):
def test_basic(self):
self.open('https://xkcd.com/1117/')
self.assert_element('img[alt="My Sky"]')
self.create_shepherd_tour()
self.add_tour_step("Welcome to XKCD!")
self.add_tour_step("This is the XKCD logo.", "#masthead img")
self.add_tour_step("Here's the daily webcomic.", "#comic img")
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
self.add_tour_step("Click here for the license.", 'a[rel="license"]')
self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]')
self.add_tour_step("Thanks for taking this tour!")
self.export_tour(filename="xkcd_tour.js") # This exports the tour
self.play_tour() # This plays the tour
| # ... existing code ...
self.add_tour_step("This is the title.", "#ctitle", alignment="top")
self.add_tour_step("Click here for the next comic.", 'a[rel="next"]')
self.add_tour_step("Click here for the previous one.", 'a[rel="prev"]')
self.add_tour_step("Learn about the author here.", 'a[rel="author"]')
self.add_tour_step("Click here for the license.", 'a[rel="license"]')
self.add_tour_step("Click for a random comic.", 'a[href*="/random/"]')
self.add_tour_step("Thanks for taking this tour!")
self.export_tour(filename="xkcd_tour.js") # This exports the tour
self.play_tour() # This plays the tour
# ... rest of the code ... |
5841f314636ee534342aa3e4530cc3ee933a052b | src/ezweb/compressor_filters.py | src/ezweb/compressor_filters.py |
from compressor.filters import FilterBase
class JSUseStrictFilter(FilterBase):
def output(self, **kwargs):
return self.remove_use_strict(self.content)
def remove_use_strict(js):
js = js.replace("'use strict';", '')
js = js.replace('"use strict";', '')
return js
|
from compressor.filters import FilterBase
class JSUseStrictFilter(FilterBase):
def output(self, **kwargs):
return self.remove_use_strict(self.content)
def remove_use_strict(self, js):
# Replacing by a ';' is safer than replacing by ''
js = js.replace("'use strict';", ';')
js = js.replace('"use strict";', ';')
return js
| Fix a bug while replacing "use strict" JS pragmas | Fix a bug while replacing "use strict" JS pragmas
| Python | agpl-3.0 | jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,rockneurotiko/wirecloud |
from compressor.filters import FilterBase
class JSUseStrictFilter(FilterBase):
def output(self, **kwargs):
return self.remove_use_strict(self.content)
- def remove_use_strict(js):
+ def remove_use_strict(self, js):
+ # Replacing by a ';' is safer than replacing by ''
- js = js.replace("'use strict';", '')
+ js = js.replace("'use strict';", ';')
- js = js.replace('"use strict";', '')
+ js = js.replace('"use strict";', ';')
return js
| Fix a bug while replacing "use strict" JS pragmas | ## Code Before:
from compressor.filters import FilterBase
class JSUseStrictFilter(FilterBase):
def output(self, **kwargs):
return self.remove_use_strict(self.content)
def remove_use_strict(js):
js = js.replace("'use strict';", '')
js = js.replace('"use strict";', '')
return js
## Instruction:
Fix a bug while replacing "use strict" JS pragmas
## Code After:
from compressor.filters import FilterBase
class JSUseStrictFilter(FilterBase):
def output(self, **kwargs):
return self.remove_use_strict(self.content)
def remove_use_strict(self, js):
# Replacing by a ';' is safer than replacing by ''
js = js.replace("'use strict';", ';')
js = js.replace('"use strict";', ';')
return js
| // ... existing code ...
return self.remove_use_strict(self.content)
def remove_use_strict(self, js):
# Replacing by a ';' is safer than replacing by ''
js = js.replace("'use strict';", ';')
js = js.replace('"use strict";', ';')
return js
// ... rest of the code ... |
7a552161eab19d24b7b221635e51a915adff0166 | templater.py | templater.py | import string
if __name__ == "__main__":
import sys
template_file = sys.argv[1]
with open(template_file) as f:
data = f.read()
template = string.Template(data)
template_mapping = {}
for item in sys.argv[2:]:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
key, value = item.split("=", 1)
template_mapping[key] = value
print("-> Using mapping: %s" % str(template_mapping))
result = template.substitute(template_mapping)
print("-----\n")
print(result)
| import string
import os
if __name__ == "__main__":
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-t", "--template", dest="template_file",
help="Input template file")
(options, args) = parser.parse_args()
if not os.path.isfile(options.template_file):
sys.stderr.write("Invalid input template file")
exit(1)
with open(options.template_file) as f:
data = f.read()
template = string.Template(data)
template_mapping = {}
for item in args:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
key, value = item.split("=", 1)
template_mapping[key] = value
print("-> Using mapping: %s" % str(template_mapping))
result = template.substitute(template_mapping)
print("-----\n")
print(result)
| Use OptionParser instead of simple sys.argv. | Use OptionParser instead of simple sys.argv.
| Python | mit | elecro/strep | import string
+ import os
+
if __name__ == "__main__":
- import sys
+ from optparse import OptionParser
- template_file = sys.argv[1]
+ parser = OptionParser()
+ parser.add_option("-t", "--template", dest="template_file",
+ help="Input template file")
+ (options, args) = parser.parse_args()
+
+ if not os.path.isfile(options.template_file):
+ sys.stderr.write("Invalid input template file")
+ exit(1)
+
- with open(template_file) as f:
+ with open(options.template_file) as f:
data = f.read()
template = string.Template(data)
template_mapping = {}
- for item in sys.argv[2:]:
+ for item in args:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
key, value = item.split("=", 1)
template_mapping[key] = value
print("-> Using mapping: %s" % str(template_mapping))
result = template.substitute(template_mapping)
print("-----\n")
print(result)
| Use OptionParser instead of simple sys.argv. | ## Code Before:
import string
if __name__ == "__main__":
import sys
template_file = sys.argv[1]
with open(template_file) as f:
data = f.read()
template = string.Template(data)
template_mapping = {}
for item in sys.argv[2:]:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
key, value = item.split("=", 1)
template_mapping[key] = value
print("-> Using mapping: %s" % str(template_mapping))
result = template.substitute(template_mapping)
print("-----\n")
print(result)
## Instruction:
Use OptionParser instead of simple sys.argv.
## Code After:
import string
import os
if __name__ == "__main__":
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-t", "--template", dest="template_file",
help="Input template file")
(options, args) = parser.parse_args()
if not os.path.isfile(options.template_file):
sys.stderr.write("Invalid input template file")
exit(1)
with open(options.template_file) as f:
data = f.read()
template = string.Template(data)
template_mapping = {}
for item in args:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
key, value = item.split("=", 1)
template_mapping[key] = value
print("-> Using mapping: %s" % str(template_mapping))
result = template.substitute(template_mapping)
print("-----\n")
print(result)
| ...
import string
import os
if __name__ == "__main__":
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-t", "--template", dest="template_file",
help="Input template file")
(options, args) = parser.parse_args()
if not os.path.isfile(options.template_file):
sys.stderr.write("Invalid input template file")
exit(1)
with open(options.template_file) as f:
data = f.read()
...
template_mapping = {}
for item in args:
# item is in the following form: KEY=VALUE
print("-> Current replacer %s" % item)
... |
6decf1f48e56832b1d15d3fc26d92f9813d13353 | coop_cms/moves.py | coop_cms/moves.py |
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
|
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| Fix HTMLParser compatibility in Python 3 | Fix HTMLParser compatibility in Python 3
| Python | bsd-3-clause | ljean/coop_cms,ljean/coop_cms,ljean/coop_cms |
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
+
+ from StringIO import StringIO
+
from HTMLParser import HTMLParser
- from StringIO import StringIO
+
+
+
else:
# Python 3
- from html.parser import HTMLParser
from io import BytesIO as StringIO
+
+ from html.parser import HTMLParser as BaseHTMLParser
+
+ class HTMLParser(BaseHTMLParser):
+ def __init__(self):
+ BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| Fix HTMLParser compatibility in Python 3 | ## Code Before:
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
## Instruction:
Fix HTMLParser compatibility in Python 3
## Code After:
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| ...
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
... |
cf8b49edfc38a98b4f6beba66bedcc13298eb114 | yunity/utils/tests/mock.py | yunity/utils/tests/mock.py | from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory
from yunity.models import Category
from yunity.utils.tests.fake import faker
class Mock(DjangoModelFactory):
class Meta:
strategy = CREATE_STRATEGY
model = None
abstract = True
class MockCategory(Mock):
class Meta:
model = "yunity.Category"
strategy = CREATE_STRATEGY
class MockUser(Mock):
class Meta:
model = "yunity.User"
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
type = Category.objects.get(name='user.default')
display_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = LazyAttribute(lambda _: faker.password())
locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)])
class MockChat(Mock):
class Meta:
model = "yunity.Chat"
strategy = CREATE_STRATEGY
administrated_by = SubFactory(MockUser)
@post_generation
def participants(self, create, extracted, **kwargs):
if not create:
return
if extracted:
for participant in extracted:
self.participants.add(participant)
| from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory
from yunity.models import Category
from yunity.utils.tests.fake import faker
class Mock(DjangoModelFactory):
class Meta:
strategy = CREATE_STRATEGY
model = None
abstract = True
class MockCategory(Mock):
class Meta:
model = "yunity.Category"
strategy = CREATE_STRATEGY
class MockUser(Mock):
class Meta:
model = "yunity.User"
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
type = Category.objects.get(name='user.default')
display_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = LazyAttribute(lambda _: faker.password())
locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)])
class MockChat(Mock):
class Meta:
model = "yunity.Chat"
strategy = CREATE_STRATEGY
administrated_by = SubFactory(MockUser)
@post_generation
def participants(self, created, participants, **kwargs):
if not created:
return
if participants:
for participant in participants:
self.participants.add(participant)
| Rename some variables to try to explain magic | Rename some variables to try to explain magic
| Python | agpl-3.0 | yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/foodsaving-backend,yunity/yunity-core | from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory
from yunity.models import Category
from yunity.utils.tests.fake import faker
class Mock(DjangoModelFactory):
class Meta:
strategy = CREATE_STRATEGY
model = None
abstract = True
class MockCategory(Mock):
class Meta:
model = "yunity.Category"
strategy = CREATE_STRATEGY
class MockUser(Mock):
class Meta:
model = "yunity.User"
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
type = Category.objects.get(name='user.default')
display_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = LazyAttribute(lambda _: faker.password())
locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)])
class MockChat(Mock):
class Meta:
model = "yunity.Chat"
strategy = CREATE_STRATEGY
administrated_by = SubFactory(MockUser)
@post_generation
- def participants(self, create, extracted, **kwargs):
+ def participants(self, created, participants, **kwargs):
- if not create:
+ if not created:
return
- if extracted:
+ if participants:
- for participant in extracted:
+ for participant in participants:
self.participants.add(participant)
| Rename some variables to try to explain magic | ## Code Before:
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory
from yunity.models import Category
from yunity.utils.tests.fake import faker
class Mock(DjangoModelFactory):
class Meta:
strategy = CREATE_STRATEGY
model = None
abstract = True
class MockCategory(Mock):
class Meta:
model = "yunity.Category"
strategy = CREATE_STRATEGY
class MockUser(Mock):
class Meta:
model = "yunity.User"
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
type = Category.objects.get(name='user.default')
display_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = LazyAttribute(lambda _: faker.password())
locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)])
class MockChat(Mock):
class Meta:
model = "yunity.Chat"
strategy = CREATE_STRATEGY
administrated_by = SubFactory(MockUser)
@post_generation
def participants(self, create, extracted, **kwargs):
if not create:
return
if extracted:
for participant in extracted:
self.participants.add(participant)
## Instruction:
Rename some variables to try to explain magic
## Code After:
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, post_generation, SubFactory
from yunity.models import Category
from yunity.utils.tests.fake import faker
class Mock(DjangoModelFactory):
class Meta:
strategy = CREATE_STRATEGY
model = None
abstract = True
class MockCategory(Mock):
class Meta:
model = "yunity.Category"
strategy = CREATE_STRATEGY
class MockUser(Mock):
class Meta:
model = "yunity.User"
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
type = Category.objects.get(name='user.default')
display_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = LazyAttribute(lambda _: faker.password())
locations = LazyAttribute(lambda _: [faker.location() for _ in range(2)])
class MockChat(Mock):
class Meta:
model = "yunity.Chat"
strategy = CREATE_STRATEGY
administrated_by = SubFactory(MockUser)
@post_generation
def participants(self, created, participants, **kwargs):
if not created:
return
if participants:
for participant in participants:
self.participants.add(participant)
| # ... existing code ...
@post_generation
def participants(self, created, participants, **kwargs):
if not created:
return
if participants:
for participant in participants:
self.participants.add(participant)
# ... rest of the code ... |
f024e340a6a443bb765b67bbdb811fa44fd3d19b | tests/test_resources.py | tests/test_resources.py | from flask import json
from helper import TestCase
from models import db, Major
class StudentsTestCase(TestCase):
def setUp(self):
super(StudentsTestCase, self).setUp()
with self.appx.app_context():
db.session.add(Major(id=1, university_id=1, name='Major1'))
db.session.add(Major(id=2, university_id=1, name='Major2'))
db.session.commit()
def test_students_patch(self):
headers = {
'Authorization': 'Bearer ' + self.jwt,
'Content-Type': 'application/json'
}
data = {
'graduation_year': 2018,
'gender': 'm',
'majors': [1, 2]
}
rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data))
self.assertEqual(rv.status_code, 200)
| from flask import json
from helper import TestCase
from models import db, Major, Student
class StudentsTestCase(TestCase):
def setUp(self):
super(StudentsTestCase, self).setUp()
with self.appx.app_context():
db.session.add(Major(id=1, university_id=1, name='Major1'))
db.session.add(Major(id=2, university_id=1, name='Major2'))
db.session.commit()
def test_students_patch(self):
headers = {
'Authorization': 'Bearer ' + self.jwt,
'Content-Type': 'application/json'
}
data = {
'graduation_year': 2018,
'gender': 'm',
'majors': [1, 2]
}
rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data))
self.assertEqual(rv.status_code, 200)
with self.appx.app_context():
student = Student.query.get(0)
self.assertEqual(student.graduation_year, data['graduation_year'])
self.assertEqual(student.gender, data['gender'])
self.assertEqual(student.majors_list, data['majors'])
| Improve testing of student patching | Improve testing of student patching
| Python | agpl-3.0 | SCUEvals/scuevals-api,SCUEvals/scuevals-api | from flask import json
from helper import TestCase
- from models import db, Major
+ from models import db, Major, Student
class StudentsTestCase(TestCase):
def setUp(self):
super(StudentsTestCase, self).setUp()
with self.appx.app_context():
db.session.add(Major(id=1, university_id=1, name='Major1'))
db.session.add(Major(id=2, university_id=1, name='Major2'))
db.session.commit()
def test_students_patch(self):
headers = {
'Authorization': 'Bearer ' + self.jwt,
'Content-Type': 'application/json'
}
data = {
'graduation_year': 2018,
'gender': 'm',
'majors': [1, 2]
}
rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data))
self.assertEqual(rv.status_code, 200)
+ with self.appx.app_context():
+ student = Student.query.get(0)
+ self.assertEqual(student.graduation_year, data['graduation_year'])
+ self.assertEqual(student.gender, data['gender'])
+ self.assertEqual(student.majors_list, data['majors'])
+ | Improve testing of student patching | ## Code Before:
from flask import json
from helper import TestCase
from models import db, Major
class StudentsTestCase(TestCase):
def setUp(self):
super(StudentsTestCase, self).setUp()
with self.appx.app_context():
db.session.add(Major(id=1, university_id=1, name='Major1'))
db.session.add(Major(id=2, university_id=1, name='Major2'))
db.session.commit()
def test_students_patch(self):
headers = {
'Authorization': 'Bearer ' + self.jwt,
'Content-Type': 'application/json'
}
data = {
'graduation_year': 2018,
'gender': 'm',
'majors': [1, 2]
}
rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data))
self.assertEqual(rv.status_code, 200)
## Instruction:
Improve testing of student patching
## Code After:
from flask import json
from helper import TestCase
from models import db, Major, Student
class StudentsTestCase(TestCase):
def setUp(self):
super(StudentsTestCase, self).setUp()
with self.appx.app_context():
db.session.add(Major(id=1, university_id=1, name='Major1'))
db.session.add(Major(id=2, university_id=1, name='Major2'))
db.session.commit()
def test_students_patch(self):
headers = {
'Authorization': 'Bearer ' + self.jwt,
'Content-Type': 'application/json'
}
data = {
'graduation_year': 2018,
'gender': 'm',
'majors': [1, 2]
}
rv = self.app.patch('/students/0', headers=headers, data=json.dumps(data))
self.assertEqual(rv.status_code, 200)
with self.appx.app_context():
student = Student.query.get(0)
self.assertEqual(student.graduation_year, data['graduation_year'])
self.assertEqual(student.gender, data['gender'])
self.assertEqual(student.majors_list, data['majors'])
| ...
from flask import json
from helper import TestCase
from models import db, Major, Student
...
self.assertEqual(rv.status_code, 200)
with self.appx.app_context():
student = Student.query.get(0)
self.assertEqual(student.graduation_year, data['graduation_year'])
self.assertEqual(student.gender, data['gender'])
self.assertEqual(student.majors_list, data['majors'])
... |
3d64eb4a7438b6b4f46f1fdf7f47d530cb11b09c | spacy/tests/regression/test_issue2396.py | spacy/tests/regression/test_issue2396.py | from __future__ import unicode_literals
from ..util import get_doc
import pytest
import numpy
@pytest.mark.parametrize('sentence,matrix', [
(
'She created a test for spacy',
numpy.array([
[0, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1],
[1, 1, 2, 3, 3, 3],
[1, 1, 3, 3, 3, 3],
[1, 1, 3, 3, 4, 4],
[1, 1, 3, 3, 4, 5]], dtype=numpy.int32)
)
])
def test_issue2396(EN, sentence, matrix):
doc = EN(sentence)
span = doc[:]
assert (doc.get_lca_matrix() == matrix).all()
assert (span.get_lca_matrix() == matrix).all()
| from __future__ import unicode_literals
from ..util import get_doc
import pytest
import numpy
from numpy.testing import assert_array_equal
@pytest.mark.parametrize('words,heads,matrix', [
(
'She created a test for spacy'.split(),
[1, 0, 1, -2, -1, -1],
numpy.array([
[0, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1],
[1, 1, 2, 3, 3, 3],
[1, 1, 3, 3, 3, 3],
[1, 1, 3, 3, 4, 4],
[1, 1, 3, 3, 4, 5]], dtype=numpy.int32)
)
])
def test_issue2396(en_vocab, words, heads, matrix):
doc = get_doc(en_vocab, words=words, heads=heads)
span = doc[:]
assert_array_equal(doc.get_lca_matrix(), matrix)
assert_array_equal(span.get_lca_matrix(), matrix)
| Update get_lca_matrix test for develop | Update get_lca_matrix test for develop
| Python | mit | explosion/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,honnibal/spaCy,honnibal/spaCy,honnibal/spaCy,spacy-io/spaCy,honnibal/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy | from __future__ import unicode_literals
from ..util import get_doc
import pytest
import numpy
+ from numpy.testing import assert_array_equal
+
- @pytest.mark.parametrize('sentence,matrix', [
+ @pytest.mark.parametrize('words,heads,matrix', [
(
- 'She created a test for spacy',
+ 'She created a test for spacy'.split(),
+ [1, 0, 1, -2, -1, -1],
numpy.array([
[0, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1],
[1, 1, 2, 3, 3, 3],
[1, 1, 3, 3, 3, 3],
[1, 1, 3, 3, 4, 4],
[1, 1, 3, 3, 4, 5]], dtype=numpy.int32)
)
])
- def test_issue2396(EN, sentence, matrix):
- doc = EN(sentence)
+ def test_issue2396(en_vocab, words, heads, matrix):
+ doc = get_doc(en_vocab, words=words, heads=heads)
+
span = doc[:]
- assert (doc.get_lca_matrix() == matrix).all()
+ assert_array_equal(doc.get_lca_matrix(), matrix)
- assert (span.get_lca_matrix() == matrix).all()
+ assert_array_equal(span.get_lca_matrix(), matrix)
| Update get_lca_matrix test for develop | ## Code Before:
from __future__ import unicode_literals
from ..util import get_doc
import pytest
import numpy
@pytest.mark.parametrize('sentence,matrix', [
(
'She created a test for spacy',
numpy.array([
[0, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1],
[1, 1, 2, 3, 3, 3],
[1, 1, 3, 3, 3, 3],
[1, 1, 3, 3, 4, 4],
[1, 1, 3, 3, 4, 5]], dtype=numpy.int32)
)
])
def test_issue2396(EN, sentence, matrix):
doc = EN(sentence)
span = doc[:]
assert (doc.get_lca_matrix() == matrix).all()
assert (span.get_lca_matrix() == matrix).all()
## Instruction:
Update get_lca_matrix test for develop
## Code After:
from __future__ import unicode_literals
from ..util import get_doc
import pytest
import numpy
from numpy.testing import assert_array_equal
@pytest.mark.parametrize('words,heads,matrix', [
(
'She created a test for spacy'.split(),
[1, 0, 1, -2, -1, -1],
numpy.array([
[0, 1, 1, 1, 1, 1],
[1, 1, 1, 1, 1, 1],
[1, 1, 2, 3, 3, 3],
[1, 1, 3, 3, 3, 3],
[1, 1, 3, 3, 4, 4],
[1, 1, 3, 3, 4, 5]], dtype=numpy.int32)
)
])
def test_issue2396(en_vocab, words, heads, matrix):
doc = get_doc(en_vocab, words=words, heads=heads)
span = doc[:]
assert_array_equal(doc.get_lca_matrix(), matrix)
assert_array_equal(span.get_lca_matrix(), matrix)
| # ... existing code ...
import pytest
import numpy
from numpy.testing import assert_array_equal
@pytest.mark.parametrize('words,heads,matrix', [
(
'She created a test for spacy'.split(),
[1, 0, 1, -2, -1, -1],
numpy.array([
[0, 1, 1, 1, 1, 1],
# ... modified code ...
)
])
def test_issue2396(en_vocab, words, heads, matrix):
doc = get_doc(en_vocab, words=words, heads=heads)
span = doc[:]
assert_array_equal(doc.get_lca_matrix(), matrix)
assert_array_equal(span.get_lca_matrix(), matrix)
# ... rest of the code ... |
210c7b7fb421a7c083b9d292370b15c0ece17fa7 | source/bark/__init__.py | source/bark/__init__.py |
from .handler.distribute import Distribute
#: Top level handler responsible for relaying all logs to other handlers.
handle = Distribute()
|
from .handler.distribute import Distribute
#: Top level handler responsible for relaying all logs to other handlers.
handler = Distribute()
handlers = handler.handlers
handle = handler.handle
| Correct handler reference variable name and add convenient accessors. | Correct handler reference variable name and add convenient accessors.
| Python | apache-2.0 | 4degrees/mill,4degrees/sawmill |
from .handler.distribute import Distribute
#: Top level handler responsible for relaying all logs to other handlers.
- handle = Distribute()
+ handler = Distribute()
+ handlers = handler.handlers
+ handle = handler.handle
+ | Correct handler reference variable name and add convenient accessors. | ## Code Before:
from .handler.distribute import Distribute
#: Top level handler responsible for relaying all logs to other handlers.
handle = Distribute()
## Instruction:
Correct handler reference variable name and add convenient accessors.
## Code After:
from .handler.distribute import Distribute
#: Top level handler responsible for relaying all logs to other handlers.
handler = Distribute()
handlers = handler.handlers
handle = handler.handle
| ...
#: Top level handler responsible for relaying all logs to other handlers.
handler = Distribute()
handlers = handler.handlers
handle = handler.handle
... |
e2722385831a0930765d2c4bb78a582d41f4b64b | src/sentry/replays.py | src/sentry/replays.py | from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
| from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| Use requests instead of httplib to do replay | Use requests instead of httplib to do replay
| Python | bsd-3-clause | beeftornado/sentry,nicholasserra/sentry,Kryz/sentry,JackDanger/sentry,imankulov/sentry,JamesMura/sentry,zenefits/sentry,kevinlondon/sentry,mvaled/sentry,JamesMura/sentry,ifduyue/sentry,looker/sentry,daevaorn/sentry,fotinakis/sentry,gencer/sentry,looker/sentry,JackDanger/sentry,mvaled/sentry,Natim/sentry,beeftornado/sentry,korealerts1/sentry,imankulov/sentry,zenefits/sentry,jean/sentry,alexm92/sentry,fotinakis/sentry,daevaorn/sentry,beeftornado/sentry,BuildingLink/sentry,ngonzalvez/sentry,BayanGroup/sentry,mvaled/sentry,mitsuhiko/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,korealerts1/sentry,mvaled/sentry,gencer/sentry,imankulov/sentry,Kryz/sentry,looker/sentry,felixbuenemann/sentry,jean/sentry,mitsuhiko/sentry,fotinakis/sentry,ifduyue/sentry,mvaled/sentry,korealerts1/sentry,kevinlondon/sentry,Natim/sentry,alexm92/sentry,zenefits/sentry,zenefits/sentry,BayanGroup/sentry,jean/sentry,mvaled/sentry,ifduyue/sentry,JamesMura/sentry,kevinlondon/sentry,BuildingLink/sentry,BayanGroup/sentry,JamesMura/sentry,nicholasserra/sentry,felixbuenemann/sentry,Kryz/sentry,BuildingLink/sentry,BuildingLink/sentry,Natim/sentry,BuildingLink/sentry,jean/sentry,JackDanger/sentry,felixbuenemann/sentry,fotinakis/sentry,nicholasserra/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,alexm92/sentry,gencer/sentry,gencer/sentry,gencer/sentry,zenefits/sentry,JamesMura/sentry,ifduyue/sentry,jean/sentry,ifduyue/sentry | from __future__ import absolute_import
+ import requests
-
- import socket
-
- from httplib import HTTPConnection, HTTPSConnection
- from urllib import urlencode
- from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
- urlparts = urlparse(self.url)
- if urlparts.scheme == 'http':
- conn_cls = HTTPConnection
- elif urlparts.scheme == 'https':
- conn_cls = HTTPSConnection
- else:
- raise ValueError(self.url)
-
- data = self.data
- if isinstance(data, dict):
- data = urlencode(data)
-
- if urlparts.query:
- full_url = urlparts.path + '?' + urlparts.query
- else:
- full_url = urlparts.path
-
- conn = conn_cls(urlparts.netloc)
try:
- conn.request(self.method, full_url, data, self.headers or {})
-
- response = conn.getresponse()
- except socket.error as e:
+ response = requests.request(
+ self.method,
+ self.url,
+ data=self.data,
+ headers=self.headers or {}
+ )
+ except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
- 'status': response.status,
+ 'status': response.status_code,
'reason': response.reason,
- 'headers': response.getheaders(),
+ 'headers': response.headers,
- 'body': response.read(),
+ 'body': response.content,
}
| Use requests instead of httplib to do replay | ## Code Before:
from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
## Instruction:
Use requests instead of httplib to do replay
## Code After:
from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| // ... existing code ...
from __future__ import absolute_import
import requests
// ... modified code ...
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
...
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
// ... rest of the code ... |
3289027d2cc5b07a83dca422bfc14114854618f8 | kazoo/__init__.py | kazoo/__init__.py | import os
from kazoo.zkclient import ZooKeeperClient
__all__ = ['ZooKeeperClient']
# ZK C client likes to spew log info to STDERR. disable that unless an
# env is present.
def disable_zookeeper_log():
import zookeeper
zookeeper.set_log_stream(open('/dev/null'))
if not "KAZOO_LOG_ENABLED" in os.environ:
disable_zookeeper_log()
def patch_extras():
# workaround for http://code.google.com/p/gevent/issues/detail?id=112
# gevent isn't patching threading._sleep which causes problems
# for Condition objects
from gevent import sleep
import threading
threading._sleep = sleep
if "KAZOO_TEST_GEVENT_PATCH" in os.environ:
from gevent import monkey; monkey.patch_all()
patch_extras()
| import os
from kazoo.zkclient import ZooKeeperClient
from kazoo.client import KazooClient
__all__ = ['ZooKeeperClient', 'KazooClient']
# ZK C client likes to spew log info to STDERR. disable that unless an
# env is present.
def disable_zookeeper_log():
import zookeeper
zookeeper.set_log_stream(open('/dev/null'))
if not "KAZOO_LOG_ENABLED" in os.environ:
disable_zookeeper_log()
def patch_extras():
# workaround for http://code.google.com/p/gevent/issues/detail?id=112
# gevent isn't patching threading._sleep which causes problems
# for Condition objects
from gevent import sleep
import threading
threading._sleep = sleep
if "KAZOO_TEST_GEVENT_PATCH" in os.environ:
from gevent import monkey; monkey.patch_all()
patch_extras()
| Add KazooClient to top-level module | Add KazooClient to top-level module | Python | apache-2.0 | nimbusproject/kazoo | import os
from kazoo.zkclient import ZooKeeperClient
+ from kazoo.client import KazooClient
- __all__ = ['ZooKeeperClient']
+ __all__ = ['ZooKeeperClient', 'KazooClient']
# ZK C client likes to spew log info to STDERR. disable that unless an
# env is present.
def disable_zookeeper_log():
import zookeeper
zookeeper.set_log_stream(open('/dev/null'))
if not "KAZOO_LOG_ENABLED" in os.environ:
disable_zookeeper_log()
def patch_extras():
# workaround for http://code.google.com/p/gevent/issues/detail?id=112
# gevent isn't patching threading._sleep which causes problems
# for Condition objects
from gevent import sleep
import threading
threading._sleep = sleep
if "KAZOO_TEST_GEVENT_PATCH" in os.environ:
from gevent import monkey; monkey.patch_all()
patch_extras()
| Add KazooClient to top-level module | ## Code Before:
import os
from kazoo.zkclient import ZooKeeperClient
__all__ = ['ZooKeeperClient']
# ZK C client likes to spew log info to STDERR. disable that unless an
# env is present.
def disable_zookeeper_log():
import zookeeper
zookeeper.set_log_stream(open('/dev/null'))
if not "KAZOO_LOG_ENABLED" in os.environ:
disable_zookeeper_log()
def patch_extras():
# workaround for http://code.google.com/p/gevent/issues/detail?id=112
# gevent isn't patching threading._sleep which causes problems
# for Condition objects
from gevent import sleep
import threading
threading._sleep = sleep
if "KAZOO_TEST_GEVENT_PATCH" in os.environ:
from gevent import monkey; monkey.patch_all()
patch_extras()
## Instruction:
Add KazooClient to top-level module
## Code After:
import os
from kazoo.zkclient import ZooKeeperClient
from kazoo.client import KazooClient
__all__ = ['ZooKeeperClient', 'KazooClient']
# ZK C client likes to spew log info to STDERR. disable that unless an
# env is present.
def disable_zookeeper_log():
import zookeeper
zookeeper.set_log_stream(open('/dev/null'))
if not "KAZOO_LOG_ENABLED" in os.environ:
disable_zookeeper_log()
def patch_extras():
# workaround for http://code.google.com/p/gevent/issues/detail?id=112
# gevent isn't patching threading._sleep which causes problems
# for Condition objects
from gevent import sleep
import threading
threading._sleep = sleep
if "KAZOO_TEST_GEVENT_PATCH" in os.environ:
from gevent import monkey; monkey.patch_all()
patch_extras()
| # ... existing code ...
from kazoo.zkclient import ZooKeeperClient
from kazoo.client import KazooClient
__all__ = ['ZooKeeperClient', 'KazooClient']
# ... rest of the code ... |
1639200e5700b1170a9d2312a32c7991ed5198b4 | tests/basics/boundmeth1.py | tests/basics/boundmeth1.py | print(type(repr([].append)))
class A:
def f(self):
return 0
def g(self, a):
return a
def h(self, a, b, c, d, e, f):
return a + b + c + d + e + f
# bound method with no extra args
m = A().f
print(m())
# bound method with 1 extra arg
m = A().g
print(m(1))
# bound method with lots of extra args
m = A().h
print(m(1, 2, 3, 4, 5, 6))
| print(type(repr([].append)))
class A:
def f(self):
return 0
def g(self, a):
return a
def h(self, a, b, c, d, e, f):
return a + b + c + d + e + f
# bound method with no extra args
m = A().f
print(m())
# bound method with 1 extra arg
m = A().g
print(m(1))
# bound method with lots of extra args
m = A().h
print(m(1, 2, 3, 4, 5, 6))
# can't assign attributes to a bound method
try:
A().f.x = 1
except AttributeError:
print('AttributeError')
| Add test for assignment of attribute to bound method. | tests/basics: Add test for assignment of attribute to bound method.
| Python | mit | ryannathans/micropython,bvernoux/micropython,HenrikSolver/micropython,dmazzella/micropython,lowRISC/micropython,toolmacher/micropython,ryannathans/micropython,cwyark/micropython,deshipu/micropython,mhoffma/micropython,HenrikSolver/micropython,Peetz0r/micropython-esp32,Timmenem/micropython,MrSurly/micropython,tralamazza/micropython,alex-robbins/micropython,chrisdearman/micropython,adafruit/circuitpython,trezor/micropython,deshipu/micropython,adafruit/circuitpython,tobbad/micropython,dmazzella/micropython,PappaPeppar/micropython,puuu/micropython,MrSurly/micropython-esp32,blazewicz/micropython,MrSurly/micropython,selste/micropython,swegener/micropython,tralamazza/micropython,mhoffma/micropython,AriZuu/micropython,PappaPeppar/micropython,lowRISC/micropython,henriknelson/micropython,torwag/micropython,puuu/micropython,toolmacher/micropython,toolmacher/micropython,kerneltask/micropython,mhoffma/micropython,deshipu/micropython,HenrikSolver/micropython,Peetz0r/micropython-esp32,mhoffma/micropython,tobbad/micropython,pozetroninc/micropython,toolmacher/micropython,AriZuu/micropython,ryannathans/micropython,hiway/micropython,pozetroninc/micropython,mhoffma/micropython,swegener/micropython,swegener/micropython,blazewicz/micropython,trezor/micropython,selste/micropython,HenrikSolver/micropython,adafruit/micropython,SHA2017-badge/micropython-esp32,henriknelson/micropython,alex-robbins/micropython,PappaPeppar/micropython,oopy/micropython,MrSurly/micropython-esp32,adafruit/micropython,TDAbboud/micropython,adafruit/circuitpython,tobbad/micropython,infinnovation/micropython,infinnovation/micropython,alex-robbins/micropython,henriknelson/micropython,pfalcon/micropython,ryannathans/micropython,adafruit/circuitpython,adafruit/circuitpython,ryannathans/micropython,MrSurly/micropython-esp32,infinnovation/micropython,TDAbboud/micropython,pozetroninc/micropython,lowRISC/micropython,adafruit/micropython,MrSurly/micropython,hiway/micropython,Peetz0r/micropython-esp32,tobbad/micropython,MrSurly/micropython,micropython/micropython-esp32,tralamazza/micropython,pramasoul/micropython,Timmenem/micropython,pfalcon/micropython,micropython/micropython-esp32,bvernoux/micropython,henriknelson/micropython,chrisdearman/micropython,adafruit/circuitpython,Timmenem/micropython,torwag/micropython,micropython/micropython-esp32,AriZuu/micropython,dmazzella/micropython,pramasoul/micropython,selste/micropython,tralamazza/micropython,lowRISC/micropython,puuu/micropython,AriZuu/micropython,SHA2017-badge/micropython-esp32,alex-robbins/micropython,adafruit/micropython,HenrikSolver/micropython,pramasoul/micropython,kerneltask/micropython,cwyark/micropython,trezor/micropython,henriknelson/micropython,tobbad/micropython,hiway/micropython,Peetz0r/micropython-esp32,PappaPeppar/micropython,AriZuu/micropython,bvernoux/micropython,TDAbboud/micropython,swegener/micropython,adafruit/micropython,selste/micropython,SHA2017-badge/micropython-esp32,Timmenem/micropython,hiway/micropython,MrSurly/micropython,infinnovation/micropython,deshipu/micropython,pozetroninc/micropython,deshipu/micropython,MrSurly/micropython-esp32,micropython/micropython-esp32,MrSurly/micropython-esp32,trezor/micropython,trezor/micropython,SHA2017-badge/micropython-esp32,torwag/micropython,pramasoul/micropython,chrisdearman/micropython,infinnovation/micropython,blazewicz/micropython,TDAbboud/micropython,pramasoul/micropython,oopy/micropython,Peetz0r/micropython-esp32,micropython/micropython-esp32,pfalcon/micropython,kerneltask/micropython,TDAbboud/micropython,chrisdearman/micropython,torwag/micropython,blazewicz/micropython,pozetroninc/micropython,cwyark/micropython,pfalcon/micropython,kerneltask/micropython,cwyark/micropython,oopy/micropython,Timmenem/micropython,toolmacher/micropython,pfalcon/micropython,puuu/micropython,SHA2017-badge/micropython-esp32,puuu/micropython,bvernoux/micropython,blazewicz/micropython,oopy/micropython,chrisdearman/micropython,kerneltask/micropython,swegener/micropython,oopy/micropython,torwag/micropython,PappaPeppar/micropython,bvernoux/micropython,alex-robbins/micropython,hiway/micropython,cwyark/micropython,selste/micropython,lowRISC/micropython,dmazzella/micropython | print(type(repr([].append)))
class A:
def f(self):
return 0
def g(self, a):
return a
def h(self, a, b, c, d, e, f):
return a + b + c + d + e + f
# bound method with no extra args
m = A().f
print(m())
# bound method with 1 extra arg
m = A().g
print(m(1))
# bound method with lots of extra args
m = A().h
print(m(1, 2, 3, 4, 5, 6))
+ # can't assign attributes to a bound method
+ try:
+ A().f.x = 1
+ except AttributeError:
+ print('AttributeError')
+ | Add test for assignment of attribute to bound method. | ## Code Before:
print(type(repr([].append)))
class A:
def f(self):
return 0
def g(self, a):
return a
def h(self, a, b, c, d, e, f):
return a + b + c + d + e + f
# bound method with no extra args
m = A().f
print(m())
# bound method with 1 extra arg
m = A().g
print(m(1))
# bound method with lots of extra args
m = A().h
print(m(1, 2, 3, 4, 5, 6))
## Instruction:
Add test for assignment of attribute to bound method.
## Code After:
print(type(repr([].append)))
class A:
def f(self):
return 0
def g(self, a):
return a
def h(self, a, b, c, d, e, f):
return a + b + c + d + e + f
# bound method with no extra args
m = A().f
print(m())
# bound method with 1 extra arg
m = A().g
print(m(1))
# bound method with lots of extra args
m = A().h
print(m(1, 2, 3, 4, 5, 6))
# can't assign attributes to a bound method
try:
A().f.x = 1
except AttributeError:
print('AttributeError')
| // ... existing code ...
m = A().h
print(m(1, 2, 3, 4, 5, 6))
# can't assign attributes to a bound method
try:
A().f.x = 1
except AttributeError:
print('AttributeError')
// ... rest of the code ... |
1e182ec0fd7cf550c809f2e6792629caeb8d5553 | sauce/lib/helpers.py | sauce/lib/helpers.py |
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
def striphtml(text):
return re.sub('<[^<]+?>', ' ', text).strip()
|
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
from genshi.core import striptags
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
| Use striptags from genshi for striphtml, since we have to have genshi anyway | Use striptags from genshi for striphtml, since we have to have genshi anyway
| Python | agpl-3.0 | moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE |
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
+ from genshi.core import striptags
+
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
+ striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
- def striphtml(text):
- return re.sub('<[^<]+?>', ' ', text).strip()
| Use striptags from genshi for striphtml, since we have to have genshi anyway | ## Code Before:
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
def striphtml(text):
return re.sub('<[^<]+?>', ' ', text).strip()
## Instruction:
Use striptags from genshi for striphtml, since we have to have genshi anyway
## Code After:
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
from genshi.core import striptags
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
| // ... existing code ...
from webhelpers.date import distance_of_time_in_words
from genshi.core import striptags
import re
// ... modified code ...
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
def link(label, url='', **attrs):
...
return link_to(label, tgurl(url), **attrs)
// ... rest of the code ... |
2ab601492a76be5d32a2e1d5009c150269e5fb03 | src/interviews/managers.py | src/interviews/managers.py | import logging
from datetime import timedelta
from django.db import models
from django.utils import timezone
from .google_analytics import get_most_read_pages
logger = logging.getLogger(__name__)
class InterviewManager(models.Manager):
def active(self, *args, **kwargs):
return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now())
def newest(self, *args, **kwargs):
return self.active().first()
def last_week(self, *args, **kwargs):
some_day_last_week = timezone.now().date() - timedelta(days=7)
monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1))
monday_of_this_week = monday_of_last_week + timedelta(days=7)
return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week,
publish__lt=monday_of_this_week)[:1]
def most_read(self, *args, **kwargs):
slugs = get_most_read_pages()
if slugs:
return self.active().filter(slug__in=slugs)
return super(InterviewManager, self).none()
| import logging
from datetime import timedelta
from django.db import models
from django.utils import timezone
from .google_analytics import get_most_read_pages
logger = logging.getLogger(__name__)
class InterviewManager(models.Manager):
def active(self, *args, **kwargs):
return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now())
def newest(self, *args, **kwargs):
return self.active().first()
def last_week(self, *args, **kwargs):
some_day_last_week = timezone.now().date() - timedelta(days=7)
monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1))
monday_of_this_week = monday_of_last_week + timedelta(days=7)
return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week,
publish__lt=monday_of_this_week)[:1]
def most_read(self, *args, **kwargs):
slugs = get_most_read_pages()
if slugs:
return self.active().filter(slug__in=slugs).order_by('slug')
return super(InterviewManager, self).none()
| Order `most_read` queryset by slug. | Order `most_read` queryset by slug.
| Python | mit | vermpy/thespotlight,vermpy/thespotlight,vermpy/thespotlight | import logging
from datetime import timedelta
from django.db import models
from django.utils import timezone
from .google_analytics import get_most_read_pages
logger = logging.getLogger(__name__)
class InterviewManager(models.Manager):
def active(self, *args, **kwargs):
return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now())
def newest(self, *args, **kwargs):
return self.active().first()
def last_week(self, *args, **kwargs):
some_day_last_week = timezone.now().date() - timedelta(days=7)
monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1))
monday_of_this_week = monday_of_last_week + timedelta(days=7)
return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week,
publish__lt=monday_of_this_week)[:1]
def most_read(self, *args, **kwargs):
slugs = get_most_read_pages()
if slugs:
- return self.active().filter(slug__in=slugs)
+ return self.active().filter(slug__in=slugs).order_by('slug')
return super(InterviewManager, self).none()
| Order `most_read` queryset by slug. | ## Code Before:
import logging
from datetime import timedelta
from django.db import models
from django.utils import timezone
from .google_analytics import get_most_read_pages
logger = logging.getLogger(__name__)
class InterviewManager(models.Manager):
def active(self, *args, **kwargs):
return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now())
def newest(self, *args, **kwargs):
return self.active().first()
def last_week(self, *args, **kwargs):
some_day_last_week = timezone.now().date() - timedelta(days=7)
monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1))
monday_of_this_week = monday_of_last_week + timedelta(days=7)
return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week,
publish__lt=monday_of_this_week)[:1]
def most_read(self, *args, **kwargs):
slugs = get_most_read_pages()
if slugs:
return self.active().filter(slug__in=slugs)
return super(InterviewManager, self).none()
## Instruction:
Order `most_read` queryset by slug.
## Code After:
import logging
from datetime import timedelta
from django.db import models
from django.utils import timezone
from .google_analytics import get_most_read_pages
logger = logging.getLogger(__name__)
class InterviewManager(models.Manager):
def active(self, *args, **kwargs):
return super(InterviewManager, self).filter(draft=False).filter(publish__lte=timezone.now())
def newest(self, *args, **kwargs):
return self.active().first()
def last_week(self, *args, **kwargs):
some_day_last_week = timezone.now().date() - timedelta(days=7)
monday_of_last_week = some_day_last_week - timedelta(days=(some_day_last_week.isocalendar()[2] - 1))
monday_of_this_week = monday_of_last_week + timedelta(days=7)
return super(InterviewManager, self).filter(draft=False).filter(publish__gte=monday_of_last_week,
publish__lt=monday_of_this_week)[:1]
def most_read(self, *args, **kwargs):
slugs = get_most_read_pages()
if slugs:
return self.active().filter(slug__in=slugs).order_by('slug')
return super(InterviewManager, self).none()
| // ... existing code ...
slugs = get_most_read_pages()
if slugs:
return self.active().filter(slug__in=slugs).order_by('slug')
return super(InterviewManager, self).none()
// ... rest of the code ... |
c37e0b66b6f0cc57d7df94f62dd47e00dc91c544 | django_archive/archivers/__init__.py | django_archive/archivers/__init__.py | from .tarball import TarballArchiver
from .zipfile import ZipArchiver
TARBALL = TarballArchiver.UNCOMPRESSED
TARBALL_GZ = TarballArchiver.GZ
TARBALL_BZ2 = TarballArchiver.BZ2
TARBALL_XZ = TarballArchiver.XZ
ZIP = 'zip'
FORMATS = (
(TARBALL, "Tarball (.tar)"),
(TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"),
(TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"),
(TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"),
(ZIP, "ZIP archive (.zip)"),
)
def get_archiver(fmt):
"""
Return the class corresponding with the provided archival format
"""
if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ):
return TarballArchiver
if fmt == ZIP:
return ZipArchiver
raise KeyError("Invalid format '{}' specified".format(fmt))
| from .tarball import TarballArchiver
from .zipfile import ZipArchiver
TARBALL = TarballArchiver.UNCOMPRESSED
TARBALL_GZ = TarballArchiver.GZ
TARBALL_BZ2 = TarballArchiver.BZ2
TARBALL_XZ = TarballArchiver.XZ
ZIP = 'zip'
FORMATS = (
TARBALL,
TARBALL_GZ,
TARBALL_BZ2,
TARBALL_XZ,
ZIP,
)
FORMATS_DESC = {
TARBALL: "Tarball (.tar)",
TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)",
TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)",
TARBALL_XZ: "xz-compressed Tarball (.tar.xz)",
ZIP: "ZIP archive (.zip)",
}
def get_archiver(fmt):
"""
Return the class corresponding with the provided archival format
"""
if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ):
return TarballArchiver
if fmt == ZIP:
return ZipArchiver
raise KeyError("Invalid format '{}' specified".format(fmt))
| Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions. | Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions.
| Python | mit | nathan-osman/django-archive,nathan-osman/django-archive | from .tarball import TarballArchiver
from .zipfile import ZipArchiver
TARBALL = TarballArchiver.UNCOMPRESSED
TARBALL_GZ = TarballArchiver.GZ
TARBALL_BZ2 = TarballArchiver.BZ2
TARBALL_XZ = TarballArchiver.XZ
ZIP = 'zip'
FORMATS = (
- (TARBALL, "Tarball (.tar)"),
- (TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"),
- (TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"),
- (TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"),
- (ZIP, "ZIP archive (.zip)"),
+ TARBALL,
+ TARBALL_GZ,
+ TARBALL_BZ2,
+ TARBALL_XZ,
+ ZIP,
)
+
+ FORMATS_DESC = {
+ TARBALL: "Tarball (.tar)",
+ TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)",
+ TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)",
+ TARBALL_XZ: "xz-compressed Tarball (.tar.xz)",
+ ZIP: "ZIP archive (.zip)",
+ }
def get_archiver(fmt):
"""
Return the class corresponding with the provided archival format
"""
if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ):
return TarballArchiver
if fmt == ZIP:
return ZipArchiver
raise KeyError("Invalid format '{}' specified".format(fmt))
| Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions. | ## Code Before:
from .tarball import TarballArchiver
from .zipfile import ZipArchiver
TARBALL = TarballArchiver.UNCOMPRESSED
TARBALL_GZ = TarballArchiver.GZ
TARBALL_BZ2 = TarballArchiver.BZ2
TARBALL_XZ = TarballArchiver.XZ
ZIP = 'zip'
FORMATS = (
(TARBALL, "Tarball (.tar)"),
(TARBALL_GZ, "gzip-compressed Tarball (.tar.gz)"),
(TARBALL_BZ2, "bzip2-compressed Tarball (.tar.bz2)"),
(TARBALL_XZ, "xz-compressed Tarball (.tar.xz)"),
(ZIP, "ZIP archive (.zip)"),
)
def get_archiver(fmt):
"""
Return the class corresponding with the provided archival format
"""
if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ):
return TarballArchiver
if fmt == ZIP:
return ZipArchiver
raise KeyError("Invalid format '{}' specified".format(fmt))
## Instruction:
Make FORMATS a tuple and add FORMATS_DESC for textual format descriptions.
## Code After:
from .tarball import TarballArchiver
from .zipfile import ZipArchiver
TARBALL = TarballArchiver.UNCOMPRESSED
TARBALL_GZ = TarballArchiver.GZ
TARBALL_BZ2 = TarballArchiver.BZ2
TARBALL_XZ = TarballArchiver.XZ
ZIP = 'zip'
FORMATS = (
TARBALL,
TARBALL_GZ,
TARBALL_BZ2,
TARBALL_XZ,
ZIP,
)
FORMATS_DESC = {
TARBALL: "Tarball (.tar)",
TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)",
TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)",
TARBALL_XZ: "xz-compressed Tarball (.tar.xz)",
ZIP: "ZIP archive (.zip)",
}
def get_archiver(fmt):
"""
Return the class corresponding with the provided archival format
"""
if fmt in (TARBALL, TARBALL_GZ, TARBALL_BZ2, TARBALL_XZ):
return TarballArchiver
if fmt == ZIP:
return ZipArchiver
raise KeyError("Invalid format '{}' specified".format(fmt))
| // ... existing code ...
FORMATS = (
TARBALL,
TARBALL_GZ,
TARBALL_BZ2,
TARBALL_XZ,
ZIP,
)
FORMATS_DESC = {
TARBALL: "Tarball (.tar)",
TARBALL_GZ: "gzip-compressed Tarball (.tar.gz)",
TARBALL_BZ2: "bzip2-compressed Tarball (.tar.bz2)",
TARBALL_XZ: "xz-compressed Tarball (.tar.xz)",
ZIP: "ZIP archive (.zip)",
}
// ... rest of the code ... |
4078743923befac99672b67ea53fd1fe11af2e8c | tests/test_mjviewer.py | tests/test_mjviewer.py | import unittest
from mujoco_py import mjviewer, mjcore
class MjLibTest(unittest.TestCase):
xml_path = 'tests/models/cartpole.xml'
def setUp(self):
self.width = 100
self.height = 100
self.viewer = mjviewer.MjViewer(visible=False,
init_width=self.width,
init_height=self.height)
def tearDown(self):
self.viewer.finish()
self.viewer = None
def test_start(self):
self.viewer.start()
self.assertTrue(self.viewer.running)
def test_render(self):
self.viewer.start()
model = mjcore.MjModel(self.xml_path)
self.viewer.set_model(model)
(data, width, height) = self.viewer.get_image()
# check image size is consistent
# note that width and height may not equal self.width and self.height
# e.g. on a computer with retina screen,
# the width and height are scaled
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
self.assertTrue(any(map(ord, data)))
| import unittest
from mujoco_py import mjviewer, mjcore
class MjLibTest(unittest.TestCase):
xml_path = 'tests/models/cartpole.xml'
def setUp(self):
self.width = 100
self.height = 100
self.viewer = mjviewer.MjViewer(visible=False,
init_width=self.width,
init_height=self.height)
def tearDown(self):
self.viewer.finish()
self.viewer = None
def test_start(self):
self.viewer.start()
self.assertTrue(self.viewer.running)
def test_render(self):
self.viewer.start()
model = mjcore.MjModel(self.xml_path)
self.viewer.set_model(model)
(data, width, height) = self.viewer.get_image()
# check image size is consistent
# note that width and height may not equal self.width and self.height
# e.g. on a computer with retina screen,
# the width and height are scaled
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
self.assertTrue(any(map(lambda x: x > 0, data)))
| Stop using ord with ints | Stop using ord with ints
| Python | mit | pulkitag/mujoco140-py,pulkitag/mujoco140-py,pulkitag/mujoco140-py | import unittest
from mujoco_py import mjviewer, mjcore
class MjLibTest(unittest.TestCase):
xml_path = 'tests/models/cartpole.xml'
def setUp(self):
self.width = 100
self.height = 100
self.viewer = mjviewer.MjViewer(visible=False,
init_width=self.width,
init_height=self.height)
def tearDown(self):
self.viewer.finish()
self.viewer = None
def test_start(self):
self.viewer.start()
self.assertTrue(self.viewer.running)
def test_render(self):
self.viewer.start()
model = mjcore.MjModel(self.xml_path)
self.viewer.set_model(model)
(data, width, height) = self.viewer.get_image()
# check image size is consistent
# note that width and height may not equal self.width and self.height
# e.g. on a computer with retina screen,
# the width and height are scaled
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
- self.assertTrue(any(map(ord, data)))
+ self.assertTrue(any(map(lambda x: x > 0, data)))
| Stop using ord with ints | ## Code Before:
import unittest
from mujoco_py import mjviewer, mjcore
class MjLibTest(unittest.TestCase):
xml_path = 'tests/models/cartpole.xml'
def setUp(self):
self.width = 100
self.height = 100
self.viewer = mjviewer.MjViewer(visible=False,
init_width=self.width,
init_height=self.height)
def tearDown(self):
self.viewer.finish()
self.viewer = None
def test_start(self):
self.viewer.start()
self.assertTrue(self.viewer.running)
def test_render(self):
self.viewer.start()
model = mjcore.MjModel(self.xml_path)
self.viewer.set_model(model)
(data, width, height) = self.viewer.get_image()
# check image size is consistent
# note that width and height may not equal self.width and self.height
# e.g. on a computer with retina screen,
# the width and height are scaled
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
self.assertTrue(any(map(ord, data)))
## Instruction:
Stop using ord with ints
## Code After:
import unittest
from mujoco_py import mjviewer, mjcore
class MjLibTest(unittest.TestCase):
xml_path = 'tests/models/cartpole.xml'
def setUp(self):
self.width = 100
self.height = 100
self.viewer = mjviewer.MjViewer(visible=False,
init_width=self.width,
init_height=self.height)
def tearDown(self):
self.viewer.finish()
self.viewer = None
def test_start(self):
self.viewer.start()
self.assertTrue(self.viewer.running)
def test_render(self):
self.viewer.start()
model = mjcore.MjModel(self.xml_path)
self.viewer.set_model(model)
(data, width, height) = self.viewer.get_image()
# check image size is consistent
# note that width and height may not equal self.width and self.height
# e.g. on a computer with retina screen,
# the width and height are scaled
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
self.assertTrue(any(map(lambda x: x > 0, data)))
| ...
self.assertEqual(len(data), 3 * width * height)
# make sure the image is not pitch black
self.assertTrue(any(map(lambda x: x > 0, data)))
... |
6ae83f01eacceb140435e72a216fa88bd97f2b0c | pyswarms/utils/console_utils.py | pyswarms/utils/console_utils.py |
""" console_utils.py: various tools for printing into console """
def cli_print(message, verbosity, threshold):
"""Helper function to print console output
Parameters
----------
message : str
the message to be printed into the console
verbosity : int
verbosity setting of the user
threshold : int
threshold for printing
"""
if verbosity >= threshold:
print(message)
else:
pass
def end_report(cost, pos, verbosity):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
Parameters
----------
cost : float
final cost from the optimization procedure.
pos : numpy.ndarray or list
best position found
verbosity : int
verbosity setting of the user.
"""
# Cuts the length of the best position if it's too long
if len(list(pos)) > 3:
out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos))
else:
out = list(pos)
template = ('================================\n'
'Optimization finished!\n'
'Final cost: {:06.4f}\n'
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
print(template) |
""" console_utils.py: various tools for printing into console """
# Import from __future__
from __future__ import with_statement
from __future__ import absolute_import
from __future__ import print_function
# Import modules
import logging
def cli_print(message, verbosity, threshold, logger):
"""Helper function to print console output
Parameters
----------
message : str
the message to be printed into the console
verbosity : int
verbosity setting of the user
threshold : int
threshold for printing
logger : logging.getLogger
logger instance
"""
if verbosity >= threshold:
logger.info(message)
else:
pass
def end_report(cost, pos, verbosity, logger):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
Parameters
----------
cost : float
final cost from the optimization procedure.
pos : numpy.ndarray or list
best position found
verbosity : int
verbosity setting of the user.
logger : logging.getLogger
logger instance
"""
# Cuts the length of the best position if it's too long
if len(list(pos)) > 3:
out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos))
else:
out = list(pos)
template = ('================================\n'
'Optimization finished!\n'
'Final cost: {:06.4f}\n'
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
logger.info(template) | Add support for logging module | Add support for logging module
This package now prints using the logging module. It can still print
onto the console, but an additional tag like INFO, DEBUG, etc. are now
being used.
Author: ljvmiranda921
| Python | mit | ljvmiranda921/pyswarms,ljvmiranda921/pyswarms |
""" console_utils.py: various tools for printing into console """
+ # Import from __future__
+ from __future__ import with_statement
+ from __future__ import absolute_import
+ from __future__ import print_function
+
+ # Import modules
+ import logging
+
- def cli_print(message, verbosity, threshold):
+ def cli_print(message, verbosity, threshold, logger):
"""Helper function to print console output
Parameters
----------
message : str
the message to be printed into the console
verbosity : int
verbosity setting of the user
threshold : int
threshold for printing
+ logger : logging.getLogger
+ logger instance
"""
if verbosity >= threshold:
- print(message)
+ logger.info(message)
else:
pass
- def end_report(cost, pos, verbosity):
+ def end_report(cost, pos, verbosity, logger):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
Parameters
----------
cost : float
final cost from the optimization procedure.
pos : numpy.ndarray or list
best position found
verbosity : int
verbosity setting of the user.
+ logger : logging.getLogger
+ logger instance
"""
# Cuts the length of the best position if it's too long
if len(list(pos)) > 3:
out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos))
else:
out = list(pos)
template = ('================================\n'
'Optimization finished!\n'
'Final cost: {:06.4f}\n'
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
- print(template)
+ logger.info(template) | Add support for logging module | ## Code Before:
""" console_utils.py: various tools for printing into console """
def cli_print(message, verbosity, threshold):
"""Helper function to print console output
Parameters
----------
message : str
the message to be printed into the console
verbosity : int
verbosity setting of the user
threshold : int
threshold for printing
"""
if verbosity >= threshold:
print(message)
else:
pass
def end_report(cost, pos, verbosity):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
Parameters
----------
cost : float
final cost from the optimization procedure.
pos : numpy.ndarray or list
best position found
verbosity : int
verbosity setting of the user.
"""
# Cuts the length of the best position if it's too long
if len(list(pos)) > 3:
out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos))
else:
out = list(pos)
template = ('================================\n'
'Optimization finished!\n'
'Final cost: {:06.4f}\n'
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
print(template)
## Instruction:
Add support for logging module
## Code After:
""" console_utils.py: various tools for printing into console """
# Import from __future__
from __future__ import with_statement
from __future__ import absolute_import
from __future__ import print_function
# Import modules
import logging
def cli_print(message, verbosity, threshold, logger):
"""Helper function to print console output
Parameters
----------
message : str
the message to be printed into the console
verbosity : int
verbosity setting of the user
threshold : int
threshold for printing
logger : logging.getLogger
logger instance
"""
if verbosity >= threshold:
logger.info(message)
else:
pass
def end_report(cost, pos, verbosity, logger):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
Parameters
----------
cost : float
final cost from the optimization procedure.
pos : numpy.ndarray or list
best position found
verbosity : int
verbosity setting of the user.
logger : logging.getLogger
logger instance
"""
# Cuts the length of the best position if it's too long
if len(list(pos)) > 3:
out = ('[ ' + 3 * '{:3f} ' + '...]').format(*list(pos))
else:
out = list(pos)
template = ('================================\n'
'Optimization finished!\n'
'Final cost: {:06.4f}\n'
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
logger.info(template) | # ... existing code ...
""" console_utils.py: various tools for printing into console """
# Import from __future__
from __future__ import with_statement
from __future__ import absolute_import
from __future__ import print_function
# Import modules
import logging
def cli_print(message, verbosity, threshold, logger):
"""Helper function to print console output
# ... modified code ...
threshold : int
threshold for printing
logger : logging.getLogger
logger instance
"""
if verbosity >= threshold:
logger.info(message)
else:
pass
def end_report(cost, pos, verbosity, logger):
"""Helper function to print a simple report at the end of the
run. This always has a threshold of 1.
...
verbosity : int
verbosity setting of the user.
logger : logging.getLogger
logger instance
"""
...
'Best value: {}\n').format(cost, out)
if verbosity >= 1:
logger.info(template)
# ... rest of the code ... |
7ed12facca2f94eb8bba721e9b11882ea24726fe | crmapp/subscribers/views.py | crmapp/subscribers/views.py | from django.shortcuts import render
from django.contrib.auth.models import User
from django.http import HttpResponseRedirect
from .forms import SubscriberForm
def subscriber_new(request, template='subscribers/subscriber_new.html'):
if request.method == 'POST':
form = SubscriberForm(request.POST)
if form.is_valid():
# Unpack form values
username = form.cleaned_data['username']
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
# Create the User record
user = User(username=username, email=email)
user.set_password(password)
user.save()
# Create Subscriber Record
# Process payment (via Stripe)
# Auto login the user
return HttpResponseRedirect('/success/')
else:
form = SubscriberForm()
return render(request, template, {'form':form})
| from django.shortcuts import render
from django.contrib.auth.models import User
from django.http import HttpResponseRedirect
from .forms import SubscriberForm
from .models import Subscriber
def subscriber_new(request, template='subscribers/subscriber_new.html'):
if request.method == 'POST':
form = SubscriberForm(request.POST)
if form.is_valid():
# Unpack form values
username = form.cleaned_data['username']
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
first_name = form.cleaned_data['first_name']
last_name = form.cleaned_data['last_name']
# Create the User record
user = User(username=username, email=email,
first_name=first_name, last_name=last_name)
user.set_password(password)
user.save()
# Create Subscriber Record
address_one = form.cleaned_data['address_one']
address_two = form.cleaned_data['address_two']
city = form.cleaned_data['city']
state = form.cleaned_data['state']
sub = Subscriber(address_one=address_one, address_two=address_two,
city=city, state=state, user_rec=user)
sub.save()
# Process payment (via Stripe)
# Auto login the user
return HttpResponseRedirect('/success/')
else:
form = SubscriberForm()
return render(request, template, {'form':form})
| Create the Subscriber Form - Part II > Update the View | Create the Subscriber Form - Part II > Update the View
| Python | mit | deenaariff/Django,tabdon/crmeasyapp,tabdon/crmeasyapp | from django.shortcuts import render
from django.contrib.auth.models import User
from django.http import HttpResponseRedirect
from .forms import SubscriberForm
+ from .models import Subscriber
def subscriber_new(request, template='subscribers/subscriber_new.html'):
if request.method == 'POST':
form = SubscriberForm(request.POST)
if form.is_valid():
# Unpack form values
username = form.cleaned_data['username']
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
+ first_name = form.cleaned_data['first_name']
+ last_name = form.cleaned_data['last_name']
# Create the User record
- user = User(username=username, email=email)
+ user = User(username=username, email=email,
+ first_name=first_name, last_name=last_name)
user.set_password(password)
user.save()
# Create Subscriber Record
+ address_one = form.cleaned_data['address_one']
+ address_two = form.cleaned_data['address_two']
+ city = form.cleaned_data['city']
+ state = form.cleaned_data['state']
+ sub = Subscriber(address_one=address_one, address_two=address_two,
+ city=city, state=state, user_rec=user)
+ sub.save()
# Process payment (via Stripe)
# Auto login the user
return HttpResponseRedirect('/success/')
else:
form = SubscriberForm()
return render(request, template, {'form':form})
| Create the Subscriber Form - Part II > Update the View | ## Code Before:
from django.shortcuts import render
from django.contrib.auth.models import User
from django.http import HttpResponseRedirect
from .forms import SubscriberForm
def subscriber_new(request, template='subscribers/subscriber_new.html'):
if request.method == 'POST':
form = SubscriberForm(request.POST)
if form.is_valid():
# Unpack form values
username = form.cleaned_data['username']
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
# Create the User record
user = User(username=username, email=email)
user.set_password(password)
user.save()
# Create Subscriber Record
# Process payment (via Stripe)
# Auto login the user
return HttpResponseRedirect('/success/')
else:
form = SubscriberForm()
return render(request, template, {'form':form})
## Instruction:
Create the Subscriber Form - Part II > Update the View
## Code After:
from django.shortcuts import render
from django.contrib.auth.models import User
from django.http import HttpResponseRedirect
from .forms import SubscriberForm
from .models import Subscriber
def subscriber_new(request, template='subscribers/subscriber_new.html'):
if request.method == 'POST':
form = SubscriberForm(request.POST)
if form.is_valid():
# Unpack form values
username = form.cleaned_data['username']
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
first_name = form.cleaned_data['first_name']
last_name = form.cleaned_data['last_name']
# Create the User record
user = User(username=username, email=email,
first_name=first_name, last_name=last_name)
user.set_password(password)
user.save()
# Create Subscriber Record
address_one = form.cleaned_data['address_one']
address_two = form.cleaned_data['address_two']
city = form.cleaned_data['city']
state = form.cleaned_data['state']
sub = Subscriber(address_one=address_one, address_two=address_two,
city=city, state=state, user_rec=user)
sub.save()
# Process payment (via Stripe)
# Auto login the user
return HttpResponseRedirect('/success/')
else:
form = SubscriberForm()
return render(request, template, {'form':form})
| // ... existing code ...
from .forms import SubscriberForm
from .models import Subscriber
def subscriber_new(request, template='subscribers/subscriber_new.html'):
// ... modified code ...
password = form.cleaned_data['password1']
email = form.cleaned_data['email']
first_name = form.cleaned_data['first_name']
last_name = form.cleaned_data['last_name']
# Create the User record
user = User(username=username, email=email,
first_name=first_name, last_name=last_name)
user.set_password(password)
user.save()
# Create Subscriber Record
address_one = form.cleaned_data['address_one']
address_two = form.cleaned_data['address_two']
city = form.cleaned_data['city']
state = form.cleaned_data['state']
sub = Subscriber(address_one=address_one, address_two=address_two,
city=city, state=state, user_rec=user)
sub.save()
# Process payment (via Stripe)
# Auto login the user
// ... rest of the code ... |
e49e7484987e3b508802adbd9e05b2b156eb6bdd | manage.py | manage.py | import os
import coverage
from flask_script import Manager, Shell
from flask_migrate import Migrate, MigrateCommand
from config import basedir
from app import create_app, db
from app.models import User, Dictionary
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
migrate = Migrate(app, db)
manager = Manager(app)
def make_shell_context():
return dict(app=app, db=db, User=User, Dictionary=Dictionary)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command("db", MigrateCommand)
cov = coverage.coverage(branch=True, include="app/*")
@manager.command
def test(coverage=False):
""" Run the unit tests. """
if coverage:
cov.start()
import unittest
tests = unittest.TestLoader().discover("tests")
unittest.TextTestRunner(verbosity=2).run(tests)
if coverage:
cov.stop()
cov.save()
print("Coverage Summary:")
cov.report()
cov_dir = os.path.join(basedir, "tmp/coverage")
cov.html_report(directory=cov_dir)
print("HTML version: %s/index.html" % cov_dir)
cov.erase()
if __name__ == "__main__":
manager.run() | import os
import coverage
from flask_script import Manager, Shell
from flask_migrate import Migrate, MigrateCommand
from config import basedir
from app import create_app, db
from app.models import User, Dictionary, Word
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
migrate = Migrate(app, db)
manager = Manager(app)
def make_shell_context():
return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command("db", MigrateCommand)
cov = coverage.coverage(branch=True, include="app/*")
@manager.command
def test(coverage=False):
""" Run the unit tests. """
if coverage:
cov.start()
import unittest
tests = unittest.TestLoader().discover("tests")
unittest.TextTestRunner(verbosity=2).run(tests)
if coverage:
cov.stop()
cov.save()
print("Coverage Summary:")
cov.report()
cov_dir = os.path.join(basedir, "tmp/coverage")
cov.html_report(directory=cov_dir)
print("HTML version: %s/index.html" % cov_dir)
cov.erase()
if __name__ == "__main__":
manager.run() | Add Word model to shell context | Add Word model to shell context
| Python | mit | Encrylize/MyDictionary,Encrylize/MyDictionary,Encrylize/MyDictionary | import os
import coverage
from flask_script import Manager, Shell
from flask_migrate import Migrate, MigrateCommand
from config import basedir
from app import create_app, db
- from app.models import User, Dictionary
+ from app.models import User, Dictionary, Word
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
migrate = Migrate(app, db)
manager = Manager(app)
def make_shell_context():
- return dict(app=app, db=db, User=User, Dictionary=Dictionary)
+ return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command("db", MigrateCommand)
cov = coverage.coverage(branch=True, include="app/*")
@manager.command
def test(coverage=False):
""" Run the unit tests. """
if coverage:
cov.start()
import unittest
tests = unittest.TestLoader().discover("tests")
unittest.TextTestRunner(verbosity=2).run(tests)
if coverage:
cov.stop()
cov.save()
print("Coverage Summary:")
cov.report()
cov_dir = os.path.join(basedir, "tmp/coverage")
cov.html_report(directory=cov_dir)
print("HTML version: %s/index.html" % cov_dir)
cov.erase()
if __name__ == "__main__":
manager.run() | Add Word model to shell context | ## Code Before:
import os
import coverage
from flask_script import Manager, Shell
from flask_migrate import Migrate, MigrateCommand
from config import basedir
from app import create_app, db
from app.models import User, Dictionary
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
migrate = Migrate(app, db)
manager = Manager(app)
def make_shell_context():
return dict(app=app, db=db, User=User, Dictionary=Dictionary)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command("db", MigrateCommand)
cov = coverage.coverage(branch=True, include="app/*")
@manager.command
def test(coverage=False):
""" Run the unit tests. """
if coverage:
cov.start()
import unittest
tests = unittest.TestLoader().discover("tests")
unittest.TextTestRunner(verbosity=2).run(tests)
if coverage:
cov.stop()
cov.save()
print("Coverage Summary:")
cov.report()
cov_dir = os.path.join(basedir, "tmp/coverage")
cov.html_report(directory=cov_dir)
print("HTML version: %s/index.html" % cov_dir)
cov.erase()
if __name__ == "__main__":
manager.run()
## Instruction:
Add Word model to shell context
## Code After:
import os
import coverage
from flask_script import Manager, Shell
from flask_migrate import Migrate, MigrateCommand
from config import basedir
from app import create_app, db
from app.models import User, Dictionary, Word
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
migrate = Migrate(app, db)
manager = Manager(app)
def make_shell_context():
return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command("db", MigrateCommand)
cov = coverage.coverage(branch=True, include="app/*")
@manager.command
def test(coverage=False):
""" Run the unit tests. """
if coverage:
cov.start()
import unittest
tests = unittest.TestLoader().discover("tests")
unittest.TextTestRunner(verbosity=2).run(tests)
if coverage:
cov.stop()
cov.save()
print("Coverage Summary:")
cov.report()
cov_dir = os.path.join(basedir, "tmp/coverage")
cov.html_report(directory=cov_dir)
print("HTML version: %s/index.html" % cov_dir)
cov.erase()
if __name__ == "__main__":
manager.run() | ...
from config import basedir
from app import create_app, db
from app.models import User, Dictionary, Word
app = create_app(os.getenv("MYDICTIONARY_CONFIG") or "default")
...
def make_shell_context():
return dict(app=app, db=db, User=User, Dictionary=Dictionary, Word=Word)
manager.add_command("shell", Shell(make_context=make_shell_context))
... |
c31c54624d7a46dfd9df96e32d2e07246868aecc | tomviz/python/DefaultITKTransform.py | tomviz/python/DefaultITKTransform.py | def transform_scalars(dataset):
"""Define this method for Python operators that
transform the input array."""
from tomviz import utils
import numpy as np
import itk
# Get the current volume as a numpy array.
array = utils.get_array(dataset)
# Set up some ITK variables
itk_image_type = itk.Image.F3
itk_converter = itk.PyBuffer[itk_image_type]
# Read the image into ITK
itk_image = itk_converter.GetImageFromArray(array)
# ITK filter (I have no idea if this is right)
filter = \
itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New()
filter.SetInitialNeighborhoodRadius(3)
filter.SetMultiplier(3)
filter.SetNumberOfIterations(25)
filter.SetReplaceValue(255)
filter.SetSeed((24,65,37))
filter.SetInput(itk_image)
filter.Update()
# Get the image back from ITK (result is a numpy image)
result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput())
# This is where the transformed data is set, it will display in tomviz.
utils.set_array(dataset, result)
| import tomviz.operators
class DefaultITKTransform(tomviz.operators.CancelableOperator):
def transform_scalars(self, dataset):
"""Define this method for Python operators that transform the input
array. This example uses an ITK filter to add 10 to each voxel value."""
# Try imports to make sure we have everything that is needed
try:
from tomviz import itkutils
import itk
except Exception as exc:
print("Could not import necessary module(s)")
raise exc
self.progress.value = 0
self.progress.maximum = 100
# Add a try/except around the ITK portion. ITK exceptions are
# passed up to the Python layer, so we can at least report what
# went wrong with the script, e.g., unsupported image type.
try:
self.progress.value = 0
self.progress.message = "Converting data to ITK image"
# Get the ITK image
itk_image = itkutils.convert_vtk_to_itk_image(dataset)
itk_input_image_type = type(itk_image)
self.progress.value = 30
self.progress.message = "Running filter"
# ITK filter
filter = itk.AddImageFilter[itk_input_image_type, # Input 1
itk_input_image_type, # Input 2
itk_input_image_type].New() # Output
filter.SetInput1(itk_image)
filter.SetConstant2(10)
itkutils.observe_filter_progress(self, filter, 30, 70)
try:
filter.Update()
except RuntimeError: # Exception thrown when ITK filter is aborted
return
self.progress.message = "Saving results"
itkutils.set_array_from_itk_image(dataset, filter.GetOutput())
self.progress.value = 100
except Exception as exc:
print("Problem encountered while running %s" %
self.__class__.__name__)
raise exc
| Change the ITK example to use a simpler ITK filter | Change the ITK example to use a simpler ITK filter
| Python | bsd-3-clause | cjh1/tomviz,cryos/tomviz,mathturtle/tomviz,OpenChemistry/tomviz,cjh1/tomviz,thewtex/tomviz,thewtex/tomviz,cryos/tomviz,mathturtle/tomviz,thewtex/tomviz,cjh1/tomviz,cryos/tomviz,OpenChemistry/tomviz,OpenChemistry/tomviz,OpenChemistry/tomviz,mathturtle/tomviz | + import tomviz.operators
- def transform_scalars(dataset):
- """Define this method for Python operators that
- transform the input array."""
- from tomviz import utils
- import numpy as np
- import itk
- # Get the current volume as a numpy array.
- array = utils.get_array(dataset)
+ class DefaultITKTransform(tomviz.operators.CancelableOperator):
- # Set up some ITK variables
- itk_image_type = itk.Image.F3
- itk_converter = itk.PyBuffer[itk_image_type]
- # Read the image into ITK
- itk_image = itk_converter.GetImageFromArray(array)
+ def transform_scalars(self, dataset):
+ """Define this method for Python operators that transform the input
+ array. This example uses an ITK filter to add 10 to each voxel value."""
+ # Try imports to make sure we have everything that is needed
+ try:
+ from tomviz import itkutils
+ import itk
+ except Exception as exc:
+ print("Could not import necessary module(s)")
+ raise exc
- # ITK filter (I have no idea if this is right)
- filter = \
- itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New()
- filter.SetInitialNeighborhoodRadius(3)
- filter.SetMultiplier(3)
- filter.SetNumberOfIterations(25)
- filter.SetReplaceValue(255)
- filter.SetSeed((24,65,37))
- filter.SetInput(itk_image)
- filter.Update()
- # Get the image back from ITK (result is a numpy image)
- result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput())
+ self.progress.value = 0
+ self.progress.maximum = 100
- # This is where the transformed data is set, it will display in tomviz.
- utils.set_array(dataset, result)
+ # Add a try/except around the ITK portion. ITK exceptions are
+ # passed up to the Python layer, so we can at least report what
+ # went wrong with the script, e.g., unsupported image type.
+ try:
+ self.progress.value = 0
+ self.progress.message = "Converting data to ITK image"
+ # Get the ITK image
+ itk_image = itkutils.convert_vtk_to_itk_image(dataset)
+ itk_input_image_type = type(itk_image)
+ self.progress.value = 30
+ self.progress.message = "Running filter"
+
+ # ITK filter
+ filter = itk.AddImageFilter[itk_input_image_type, # Input 1
+ itk_input_image_type, # Input 2
+ itk_input_image_type].New() # Output
+ filter.SetInput1(itk_image)
+ filter.SetConstant2(10)
+ itkutils.observe_filter_progress(self, filter, 30, 70)
+
+ try:
+ filter.Update()
+ except RuntimeError: # Exception thrown when ITK filter is aborted
+ return
+
+ self.progress.message = "Saving results"
+
+ itkutils.set_array_from_itk_image(dataset, filter.GetOutput())
+
+ self.progress.value = 100
+ except Exception as exc:
+ print("Problem encountered while running %s" %
+ self.__class__.__name__)
+ raise exc
+ | Change the ITK example to use a simpler ITK filter | ## Code Before:
def transform_scalars(dataset):
"""Define this method for Python operators that
transform the input array."""
from tomviz import utils
import numpy as np
import itk
# Get the current volume as a numpy array.
array = utils.get_array(dataset)
# Set up some ITK variables
itk_image_type = itk.Image.F3
itk_converter = itk.PyBuffer[itk_image_type]
# Read the image into ITK
itk_image = itk_converter.GetImageFromArray(array)
# ITK filter (I have no idea if this is right)
filter = \
itk.ConfidenceConnectedImageFilter[itk_image_type,itk.Image.SS3].New()
filter.SetInitialNeighborhoodRadius(3)
filter.SetMultiplier(3)
filter.SetNumberOfIterations(25)
filter.SetReplaceValue(255)
filter.SetSeed((24,65,37))
filter.SetInput(itk_image)
filter.Update()
# Get the image back from ITK (result is a numpy image)
result = itk.PyBuffer[itk.Image.SS3].GetArrayFromImage(filter.GetOutput())
# This is where the transformed data is set, it will display in tomviz.
utils.set_array(dataset, result)
## Instruction:
Change the ITK example to use a simpler ITK filter
## Code After:
import tomviz.operators
class DefaultITKTransform(tomviz.operators.CancelableOperator):
def transform_scalars(self, dataset):
"""Define this method for Python operators that transform the input
array. This example uses an ITK filter to add 10 to each voxel value."""
# Try imports to make sure we have everything that is needed
try:
from tomviz import itkutils
import itk
except Exception as exc:
print("Could not import necessary module(s)")
raise exc
self.progress.value = 0
self.progress.maximum = 100
# Add a try/except around the ITK portion. ITK exceptions are
# passed up to the Python layer, so we can at least report what
# went wrong with the script, e.g., unsupported image type.
try:
self.progress.value = 0
self.progress.message = "Converting data to ITK image"
# Get the ITK image
itk_image = itkutils.convert_vtk_to_itk_image(dataset)
itk_input_image_type = type(itk_image)
self.progress.value = 30
self.progress.message = "Running filter"
# ITK filter
filter = itk.AddImageFilter[itk_input_image_type, # Input 1
itk_input_image_type, # Input 2
itk_input_image_type].New() # Output
filter.SetInput1(itk_image)
filter.SetConstant2(10)
itkutils.observe_filter_progress(self, filter, 30, 70)
try:
filter.Update()
except RuntimeError: # Exception thrown when ITK filter is aborted
return
self.progress.message = "Saving results"
itkutils.set_array_from_itk_image(dataset, filter.GetOutput())
self.progress.value = 100
except Exception as exc:
print("Problem encountered while running %s" %
self.__class__.__name__)
raise exc
| // ... existing code ...
import tomviz.operators
class DefaultITKTransform(tomviz.operators.CancelableOperator):
def transform_scalars(self, dataset):
"""Define this method for Python operators that transform the input
array. This example uses an ITK filter to add 10 to each voxel value."""
# Try imports to make sure we have everything that is needed
try:
from tomviz import itkutils
import itk
except Exception as exc:
print("Could not import necessary module(s)")
raise exc
self.progress.value = 0
self.progress.maximum = 100
# Add a try/except around the ITK portion. ITK exceptions are
# passed up to the Python layer, so we can at least report what
# went wrong with the script, e.g., unsupported image type.
try:
self.progress.value = 0
self.progress.message = "Converting data to ITK image"
# Get the ITK image
itk_image = itkutils.convert_vtk_to_itk_image(dataset)
itk_input_image_type = type(itk_image)
self.progress.value = 30
self.progress.message = "Running filter"
# ITK filter
filter = itk.AddImageFilter[itk_input_image_type, # Input 1
itk_input_image_type, # Input 2
itk_input_image_type].New() # Output
filter.SetInput1(itk_image)
filter.SetConstant2(10)
itkutils.observe_filter_progress(self, filter, 30, 70)
try:
filter.Update()
except RuntimeError: # Exception thrown when ITK filter is aborted
return
self.progress.message = "Saving results"
itkutils.set_array_from_itk_image(dataset, filter.GetOutput())
self.progress.value = 100
except Exception as exc:
print("Problem encountered while running %s" %
self.__class__.__name__)
raise exc
// ... rest of the code ... |
0887e200f31edd8d61e0dd1d3fefae7e828c9269 | mindbender/maya/plugins/validate_single_assembly.py | mindbender/maya/plugins/validate_single_assembly.py | import pyblish.api
class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin):
"""Each asset must have a single top-level group
The given instance is test-exported, along with construction
history to test whether more than 1 top-level DAG node would
be included in the exported file.
"""
label = "Validate Single Assembly"
order = pyblish.api.ValidatorOrder
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
def process(self, instance):
from maya import cmds
from mindbender import maya
with maya.maintained_selection():
cmds.select(instance, replace=True)
nodes = cmds.file(
constructionHistory=True,
exportSelected=True,
preview=True,
force=True,
)
assemblies = cmds.ls(nodes, assemblies=True)
if not assemblies:
raise Exception("No assembly found.")
if len(assemblies) != 1:
assemblies = '"%s"' % '", "'.join(assemblies)
raise Exception(
"Multiple assemblies found: %s" % assemblies
)
| import pyblish.api
class SelectAssemblies(pyblish.api.Action):
label = "Select Assemblies"
on = "failed"
def process(self, context, plugin):
from maya import cmds
cmds.select(plugin.assemblies)
class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin):
"""Each asset must have a single top-level group
The given instance is test-exported, along with construction
history to test whether more than 1 top-level DAG node would
be included in the exported file.
"""
label = "Validate Single Assembly"
order = pyblish.api.ValidatorOrder
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
actions = [
pyblish.api.Category("Actions"),
SelectAssemblies,
]
assemblies = []
def process(self, instance):
from maya import cmds
from mindbender import maya
with maya.maintained_selection():
cmds.select(instance, replace=True)
nodes = cmds.file(
constructionHistory=True,
exportSelected=True,
preview=True,
force=True,
)
self.assemblies[:] = cmds.ls(nodes, assemblies=True)
if not self.assemblies:
raise Exception("No assembly found.")
if len(self.assemblies) != 1:
self.assemblies = '"%s"' % '", "'.join(self.assemblies)
raise Exception(
"Multiple assemblies found: %s" % self.assemblies
)
| Add action to select the multiple assemblies. | Add action to select the multiple assemblies.
| Python | mit | getavalon/core,MoonShineVFX/core,MoonShineVFX/core,mindbender-studio/core,mindbender-studio/core,getavalon/core | import pyblish.api
+
+
+ class SelectAssemblies(pyblish.api.Action):
+ label = "Select Assemblies"
+ on = "failed"
+
+ def process(self, context, plugin):
+ from maya import cmds
+ cmds.select(plugin.assemblies)
class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin):
"""Each asset must have a single top-level group
The given instance is test-exported, along with construction
history to test whether more than 1 top-level DAG node would
be included in the exported file.
"""
label = "Validate Single Assembly"
order = pyblish.api.ValidatorOrder
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
+ actions = [
+ pyblish.api.Category("Actions"),
+ SelectAssemblies,
+ ]
+
+ assemblies = []
def process(self, instance):
from maya import cmds
from mindbender import maya
with maya.maintained_selection():
cmds.select(instance, replace=True)
nodes = cmds.file(
constructionHistory=True,
exportSelected=True,
preview=True,
force=True,
)
- assemblies = cmds.ls(nodes, assemblies=True)
+ self.assemblies[:] = cmds.ls(nodes, assemblies=True)
- if not assemblies:
+ if not self.assemblies:
raise Exception("No assembly found.")
- if len(assemblies) != 1:
+ if len(self.assemblies) != 1:
- assemblies = '"%s"' % '", "'.join(assemblies)
+ self.assemblies = '"%s"' % '", "'.join(self.assemblies)
raise Exception(
- "Multiple assemblies found: %s" % assemblies
+ "Multiple assemblies found: %s" % self.assemblies
)
| Add action to select the multiple assemblies. | ## Code Before:
import pyblish.api
class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin):
"""Each asset must have a single top-level group
The given instance is test-exported, along with construction
history to test whether more than 1 top-level DAG node would
be included in the exported file.
"""
label = "Validate Single Assembly"
order = pyblish.api.ValidatorOrder
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
def process(self, instance):
from maya import cmds
from mindbender import maya
with maya.maintained_selection():
cmds.select(instance, replace=True)
nodes = cmds.file(
constructionHistory=True,
exportSelected=True,
preview=True,
force=True,
)
assemblies = cmds.ls(nodes, assemblies=True)
if not assemblies:
raise Exception("No assembly found.")
if len(assemblies) != 1:
assemblies = '"%s"' % '", "'.join(assemblies)
raise Exception(
"Multiple assemblies found: %s" % assemblies
)
## Instruction:
Add action to select the multiple assemblies.
## Code After:
import pyblish.api
class SelectAssemblies(pyblish.api.Action):
label = "Select Assemblies"
on = "failed"
def process(self, context, plugin):
from maya import cmds
cmds.select(plugin.assemblies)
class ValidateMindbenderSingleAssembly(pyblish.api.InstancePlugin):
"""Each asset must have a single top-level group
The given instance is test-exported, along with construction
history to test whether more than 1 top-level DAG node would
be included in the exported file.
"""
label = "Validate Single Assembly"
order = pyblish.api.ValidatorOrder
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
actions = [
pyblish.api.Category("Actions"),
SelectAssemblies,
]
assemblies = []
def process(self, instance):
from maya import cmds
from mindbender import maya
with maya.maintained_selection():
cmds.select(instance, replace=True)
nodes = cmds.file(
constructionHistory=True,
exportSelected=True,
preview=True,
force=True,
)
self.assemblies[:] = cmds.ls(nodes, assemblies=True)
if not self.assemblies:
raise Exception("No assembly found.")
if len(self.assemblies) != 1:
self.assemblies = '"%s"' % '", "'.join(self.assemblies)
raise Exception(
"Multiple assemblies found: %s" % self.assemblies
)
| # ... existing code ...
import pyblish.api
class SelectAssemblies(pyblish.api.Action):
label = "Select Assemblies"
on = "failed"
def process(self, context, plugin):
from maya import cmds
cmds.select(plugin.assemblies)
# ... modified code ...
hosts = ["maya"]
families = ["mindbender.model", "mindbender.rig"]
actions = [
pyblish.api.Category("Actions"),
SelectAssemblies,
]
assemblies = []
def process(self, instance):
...
)
self.assemblies[:] = cmds.ls(nodes, assemblies=True)
if not self.assemblies:
raise Exception("No assembly found.")
if len(self.assemblies) != 1:
self.assemblies = '"%s"' % '", "'.join(self.assemblies)
raise Exception(
"Multiple assemblies found: %s" % self.assemblies
)
# ... rest of the code ... |
b2bc77023ed3e19f6f7483645e2a11952c061de0 | tests/registryd/test_registry_startup.py | tests/registryd/test_registry_startup.py | PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
def test_accessible_iface_properties(registry, session_manager):
val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE)
assert str(val) == 'main'
| PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
def get_property(proxy, iface_name, prop_name):
return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)
def test_accessible_iface_properties(registry, session_manager):
values = [
('Name', 'main'),
('Description', ''),
]
for prop_name, expected in values:
assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
| Test the Description property of the registry's root | Test the Description property of the registry's root
| Python | lgpl-2.1 | GNOME/at-spi2-core,GNOME/at-spi2-core,GNOME/at-spi2-core | PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
+ def get_property(proxy, iface_name, prop_name):
+ return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)
+
def test_accessible_iface_properties(registry, session_manager):
- val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE)
- assert str(val) == 'main'
+ values = [
+ ('Name', 'main'),
+ ('Description', ''),
+ ]
+ for prop_name, expected in values:
+ assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
+ | Test the Description property of the registry's root | ## Code Before:
PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
def test_accessible_iface_properties(registry, session_manager):
val = registry.Get(ACCESSIBLE_IFACE, 'Name', dbus_interface=PROPERTIES_IFACE)
assert str(val) == 'main'
## Instruction:
Test the Description property of the registry's root
## Code After:
PROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
def get_property(proxy, iface_name, prop_name):
return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)
def test_accessible_iface_properties(registry, session_manager):
values = [
('Name', 'main'),
('Description', ''),
]
for prop_name, expected in values:
assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
| # ... existing code ...
ACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'
def get_property(proxy, iface_name, prop_name):
return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)
def test_accessible_iface_properties(registry, session_manager):
values = [
('Name', 'main'),
('Description', ''),
]
for prop_name, expected in values:
assert get_property(registry, ACCESSIBLE_IFACE, prop_name) == expected
# ... rest of the code ... |
805c52698b3fed8df98462c15045f5de3822e241 | edx_repo_tools/dev/clone_org.py | edx_repo_tools/dev/clone_org.py | """Clone an entire GitHub organization."""
import os.path
import click
from git.repo.base import Repo
from edx_repo_tools.auth import pass_github
@click.command()
@click.option(
'--forks/--no-forks', is_flag=True, default=False,
help="Should forks be included?"
)
@click.option(
'--depth', type=int, default=0,
help="Depth argument for git clone",
)
@click.argument(
'org'
)
@pass_github
def main(hub, forks, depth, org):
for repo in hub.organization(org).repositories():
if repo.fork and not forks:
continue
dir_name = repo.name
dir_name = dir_name.lstrip("-") # avoid dirname/option confusion
if os.path.exists(dir_name):
continue
print(repo.full_name)
clone_args = {}
if depth:
clone_args['depth'] = depth
Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
| """Clone an entire GitHub organization."""
import os.path
import click
from git.repo.base import Repo
from edx_repo_tools.auth import pass_github
@click.command()
@click.option(
'--forks/--no-forks', is_flag=True, default=False,
help="Should forks be included?"
)
@click.option(
'--depth', type=int, default=0,
help="Depth argument for git clone",
)
@click.argument(
'org'
)
@pass_github
def main(hub, forks, depth, org):
for repo in hub.organization(org).iter_repos():
if repo.fork and not forks:
continue
dir_name = repo.name
dir_name = dir_name.lstrip("-") # avoid dirname/option confusion
if os.path.exists(dir_name):
continue
print(repo.full_name)
clone_args = {}
if depth:
clone_args['depth'] = depth
Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
| Fix to work in python 3. | Fix to work in python 3.
| Python | apache-2.0 | edx/repo-tools,edx/repo-tools | """Clone an entire GitHub organization."""
import os.path
import click
from git.repo.base import Repo
from edx_repo_tools.auth import pass_github
-
@click.command()
@click.option(
'--forks/--no-forks', is_flag=True, default=False,
help="Should forks be included?"
)
@click.option(
'--depth', type=int, default=0,
help="Depth argument for git clone",
)
@click.argument(
'org'
)
@pass_github
def main(hub, forks, depth, org):
- for repo in hub.organization(org).repositories():
+ for repo in hub.organization(org).iter_repos():
if repo.fork and not forks:
continue
dir_name = repo.name
dir_name = dir_name.lstrip("-") # avoid dirname/option confusion
if os.path.exists(dir_name):
continue
print(repo.full_name)
clone_args = {}
if depth:
clone_args['depth'] = depth
Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
| Fix to work in python 3. | ## Code Before:
"""Clone an entire GitHub organization."""
import os.path
import click
from git.repo.base import Repo
from edx_repo_tools.auth import pass_github
@click.command()
@click.option(
'--forks/--no-forks', is_flag=True, default=False,
help="Should forks be included?"
)
@click.option(
'--depth', type=int, default=0,
help="Depth argument for git clone",
)
@click.argument(
'org'
)
@pass_github
def main(hub, forks, depth, org):
for repo in hub.organization(org).repositories():
if repo.fork and not forks:
continue
dir_name = repo.name
dir_name = dir_name.lstrip("-") # avoid dirname/option confusion
if os.path.exists(dir_name):
continue
print(repo.full_name)
clone_args = {}
if depth:
clone_args['depth'] = depth
Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
## Instruction:
Fix to work in python 3.
## Code After:
"""Clone an entire GitHub organization."""
import os.path
import click
from git.repo.base import Repo
from edx_repo_tools.auth import pass_github
@click.command()
@click.option(
'--forks/--no-forks', is_flag=True, default=False,
help="Should forks be included?"
)
@click.option(
'--depth', type=int, default=0,
help="Depth argument for git clone",
)
@click.argument(
'org'
)
@pass_github
def main(hub, forks, depth, org):
for repo in hub.organization(org).iter_repos():
if repo.fork and not forks:
continue
dir_name = repo.name
dir_name = dir_name.lstrip("-") # avoid dirname/option confusion
if os.path.exists(dir_name):
continue
print(repo.full_name)
clone_args = {}
if depth:
clone_args['depth'] = depth
Repo.clone_from(repo.ssh_url, dir_name, **clone_args)
| // ... existing code ...
from edx_repo_tools.auth import pass_github
@click.command()
// ... modified code ...
@pass_github
def main(hub, forks, depth, org):
for repo in hub.organization(org).iter_repos():
if repo.fork and not forks:
continue
// ... rest of the code ... |
66946f72d243f1836df0dbd8917f204011ec1701 | hs_core/autocomplete_light_registry.py | hs_core/autocomplete_light_registry.py | from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields = ['username', 'first_name', 'last_name']
split_words = True
def choices_for_request(self):
self.choices = self.choices.filter(is_active=True)
return super(UserAutocomplete, self).choices_for_request()
def choice_label(self, choice):
label = ""
if choice.first_name:
label += choice.first_name
if choice.last_name:
if choice.first_name:
label += " "
label += choice.last_name
if choice.userprofile.organization:
if choice.first_name or choice.last_name:
label += ", "
label += choice.userprofile.organization
if choice.username:
label += "".join([" (", choice.username, ")"])
return label
autocomplete_light.register(User, UserAutocomplete)
class GroupAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields=['name']
def choices_for_request(self):
self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author')
return super(GroupAutocomplete, self).choices_for_request()
autocomplete_light.register(Group, GroupAutocomplete)
| from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields = ['username', 'first_name', 'last_name']
split_words = True
def choices_for_request(self):
self.choices = self.choices.filter(is_active=True)
return super(UserAutocomplete, self).choices_for_request()
def choice_label(self, choice):
label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""])
if choice.userprofile.organization:
if choice.first_name or choice.last_name:
label += ", "
label += choice.userprofile.organization
if choice.username:
label += "".join([" (", choice.username, ")"])
return label
autocomplete_light.register(User, UserAutocomplete)
class GroupAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields=['name']
def choices_for_request(self):
self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author')
return super(GroupAutocomplete, self).choices_for_request()
autocomplete_light.register(Group, GroupAutocomplete)
| Add middle name display to autocomplete widget | Add middle name display to autocomplete widget
| Python | bsd-3-clause | hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare,hydroshare/hydroshare | from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
+
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields = ['username', 'first_name', 'last_name']
split_words = True
def choices_for_request(self):
self.choices = self.choices.filter(is_active=True)
return super(UserAutocomplete, self).choices_for_request()
def choice_label(self, choice):
+ label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""])
- label = ""
-
- if choice.first_name:
- label += choice.first_name
-
- if choice.last_name:
- if choice.first_name:
- label += " "
- label += choice.last_name
if choice.userprofile.organization:
if choice.first_name or choice.last_name:
label += ", "
label += choice.userprofile.organization
if choice.username:
label += "".join([" (", choice.username, ")"])
return label
autocomplete_light.register(User, UserAutocomplete)
class GroupAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields=['name']
def choices_for_request(self):
self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author')
return super(GroupAutocomplete, self).choices_for_request()
autocomplete_light.register(Group, GroupAutocomplete)
| Add middle name display to autocomplete widget | ## Code Before:
from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields = ['username', 'first_name', 'last_name']
split_words = True
def choices_for_request(self):
self.choices = self.choices.filter(is_active=True)
return super(UserAutocomplete, self).choices_for_request()
def choice_label(self, choice):
label = ""
if choice.first_name:
label += choice.first_name
if choice.last_name:
if choice.first_name:
label += " "
label += choice.last_name
if choice.userprofile.organization:
if choice.first_name or choice.last_name:
label += ", "
label += choice.userprofile.organization
if choice.username:
label += "".join([" (", choice.username, ")"])
return label
autocomplete_light.register(User, UserAutocomplete)
class GroupAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields=['name']
def choices_for_request(self):
self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author')
return super(GroupAutocomplete, self).choices_for_request()
autocomplete_light.register(Group, GroupAutocomplete)
## Instruction:
Add middle name display to autocomplete widget
## Code After:
from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields = ['username', 'first_name', 'last_name']
split_words = True
def choices_for_request(self):
self.choices = self.choices.filter(is_active=True)
return super(UserAutocomplete, self).choices_for_request()
def choice_label(self, choice):
label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""])
if choice.userprofile.organization:
if choice.first_name or choice.last_name:
label += ", "
label += choice.userprofile.organization
if choice.username:
label += "".join([" (", choice.username, ")"])
return label
autocomplete_light.register(User, UserAutocomplete)
class GroupAutocomplete(autocomplete_light.AutocompleteModelBase):
search_fields=['name']
def choices_for_request(self):
self.choices = self.choices.filter(gaccess__active=True).exclude(name='Hydroshare Author')
return super(GroupAutocomplete, self).choices_for_request()
autocomplete_light.register(Group, GroupAutocomplete)
| ...
from autocomplete_light import shortcuts as autocomplete_light
from django.contrib.auth.models import User, Group
class UserAutocomplete(autocomplete_light.AutocompleteModelBase):
...
def choice_label(self, choice):
label = " ".join([choice.first_name or "", choice.userprofile.middle_name or "", choice.last_name or ""])
if choice.userprofile.organization:
... |
bddab649c6684f09870983dca97c39eb30b62c06 | djangobotcfg/status.py | djangobotcfg/status.py | from buildbot.status import html, words
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
# authz = Authz(
# forceBuild=True,
# forceAllBuilds=True,
# pingBuilder=True,
# gracefulShutdown=True,
# stopBuild=True,
# stopAllBuilds=True,
# cancelPendingBuild=True,
# cleanShutdown=True,
# )
def get_status():
return [
html.WebStatus(
http_port = '8010',
# authz = authz,
order_console_by_time = True,
revlink = 'http://code.djangoproject.com/changeset/%s',
changecommentlink = (
r'\b#(\d+)\b',
r'http://code.djangoproject.com/ticket/\1',
r'Ticket \g<0>'
)
),
words.IRC(
host = 'irc.freenode.net',
channels = ['#revsys'],
nick = 'djangobuilds',
notify_events = {
'successToFailure': True,
'failureToSuccess': True,
}
)
] | from buildbot.status import html, words
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
def get_status():
return [
html.WebStatus(
http_port = '8010',
# authz = authz,
order_console_by_time = True,
revlink = 'http://code.djangoproject.com/changeset/%s',
changecommentlink = (
r'\b#(\d+)\b',
r'http://code.djangoproject.com/ticket/\1',
r'Ticket \g<0>'
)
),
] | Remove the IRC bot for now, and also the commented-out code. | Remove the IRC bot for now, and also the commented-out code.
| Python | bsd-3-clause | hochanh/django-buildmaster,jacobian-archive/django-buildmaster | from buildbot.status import html, words
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
-
- # authz = Authz(
- # forceBuild=True,
- # forceAllBuilds=True,
- # pingBuilder=True,
- # gracefulShutdown=True,
- # stopBuild=True,
- # stopAllBuilds=True,
- # cancelPendingBuild=True,
- # cleanShutdown=True,
- # )
def get_status():
return [
html.WebStatus(
http_port = '8010',
# authz = authz,
order_console_by_time = True,
revlink = 'http://code.djangoproject.com/changeset/%s',
changecommentlink = (
r'\b#(\d+)\b',
r'http://code.djangoproject.com/ticket/\1',
r'Ticket \g<0>'
)
),
-
- words.IRC(
- host = 'irc.freenode.net',
- channels = ['#revsys'],
- nick = 'djangobuilds',
- notify_events = {
- 'successToFailure': True,
- 'failureToSuccess': True,
- }
- )
] | Remove the IRC bot for now, and also the commented-out code. | ## Code Before:
from buildbot.status import html, words
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
# authz = Authz(
# forceBuild=True,
# forceAllBuilds=True,
# pingBuilder=True,
# gracefulShutdown=True,
# stopBuild=True,
# stopAllBuilds=True,
# cancelPendingBuild=True,
# cleanShutdown=True,
# )
def get_status():
return [
html.WebStatus(
http_port = '8010',
# authz = authz,
order_console_by_time = True,
revlink = 'http://code.djangoproject.com/changeset/%s',
changecommentlink = (
r'\b#(\d+)\b',
r'http://code.djangoproject.com/ticket/\1',
r'Ticket \g<0>'
)
),
words.IRC(
host = 'irc.freenode.net',
channels = ['#revsys'],
nick = 'djangobuilds',
notify_events = {
'successToFailure': True,
'failureToSuccess': True,
}
)
]
## Instruction:
Remove the IRC bot for now, and also the commented-out code.
## Code After:
from buildbot.status import html, words
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
def get_status():
return [
html.WebStatus(
http_port = '8010',
# authz = authz,
order_console_by_time = True,
revlink = 'http://code.djangoproject.com/changeset/%s',
changecommentlink = (
r'\b#(\d+)\b',
r'http://code.djangoproject.com/ticket/\1',
r'Ticket \g<0>'
)
),
] | ...
from buildbot.status.web.authz import Authz
from buildbot.status.web.auth import BasicAuth
def get_status():
...
)
),
]
... |
a4375a6ec5ca54b887527885235317986011801c | guesser.py | guesser.py | from synt.utils.redis_manager import RedisManager
from synt.utils.extractors import best_word_feats
from synt.utils.text import sanitize_text
MANAGER = RedisManager()
DEFAULT_CLASSIFIER = MANAGER.load_classifier()
def guess(text, classifier=DEFAULT_CLASSIFIER):
"""Takes a blob of text and returns the sentiment and confidence score."""
assert classifier, "Needs a classifier."
bag_of_words = best_word_feats(sanitize_text(text))
if bag_of_words:
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
return (guess, [(prob.prob(sample),sample) for sample in prob.samples()])
| from synt.utils.redis_manager import RedisManager
from synt.utils.extractors import best_word_feats
from synt.utils.text import sanitize_text
MANAGER = RedisManager()
DEFAULT_CLASSIFIER = MANAGER.load_classifier()
def guess(text, classifier=DEFAULT_CLASSIFIER):
"""Takes a blob of text and returns the sentiment and confidence score."""
assert classifier, "Needs a classifier."
bag_of_words = best_word_feats(sanitize_text(text))
if bag_of_words:
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
#return a -1 .. 1 score
score = prob.prob('positive') - prob.prob('negative')
return score
| Return a -1 .. 1 sentiment score. | Return a -1 .. 1 sentiment score.
| Python | agpl-3.0 | lrvick/synt | from synt.utils.redis_manager import RedisManager
from synt.utils.extractors import best_word_feats
from synt.utils.text import sanitize_text
MANAGER = RedisManager()
DEFAULT_CLASSIFIER = MANAGER.load_classifier()
def guess(text, classifier=DEFAULT_CLASSIFIER):
"""Takes a blob of text and returns the sentiment and confidence score."""
assert classifier, "Needs a classifier."
bag_of_words = best_word_feats(sanitize_text(text))
if bag_of_words:
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
- return (guess, [(prob.prob(sample),sample) for sample in prob.samples()])
+
+ #return a -1 .. 1 score
+ score = prob.prob('positive') - prob.prob('negative')
+
+ return score
| Return a -1 .. 1 sentiment score. | ## Code Before:
from synt.utils.redis_manager import RedisManager
from synt.utils.extractors import best_word_feats
from synt.utils.text import sanitize_text
MANAGER = RedisManager()
DEFAULT_CLASSIFIER = MANAGER.load_classifier()
def guess(text, classifier=DEFAULT_CLASSIFIER):
"""Takes a blob of text and returns the sentiment and confidence score."""
assert classifier, "Needs a classifier."
bag_of_words = best_word_feats(sanitize_text(text))
if bag_of_words:
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
return (guess, [(prob.prob(sample),sample) for sample in prob.samples()])
## Instruction:
Return a -1 .. 1 sentiment score.
## Code After:
from synt.utils.redis_manager import RedisManager
from synt.utils.extractors import best_word_feats
from synt.utils.text import sanitize_text
MANAGER = RedisManager()
DEFAULT_CLASSIFIER = MANAGER.load_classifier()
def guess(text, classifier=DEFAULT_CLASSIFIER):
"""Takes a blob of text and returns the sentiment and confidence score."""
assert classifier, "Needs a classifier."
bag_of_words = best_word_feats(sanitize_text(text))
if bag_of_words:
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
#return a -1 .. 1 score
score = prob.prob('positive') - prob.prob('negative')
return score
| # ... existing code ...
guess = classifier.classify(bag_of_words)
prob = classifier.prob_classify(bag_of_words)
#return a -1 .. 1 score
score = prob.prob('positive') - prob.prob('negative')
return score
# ... rest of the code ... |
f8bdd7c8139cfc6d7af4bb3d89e983073db976bf | mecodesktop.py | mecodesktop.py |
from macroeco import desktop
desktop()
|
import sys as _sys
from macroeco import desktop
if len(_sys.argv) > 1:
desktop(_sys.argv[1])
else:
desktop()
| Allow compiled OS X app to take parameter file as input on command line | Allow compiled OS X app to take parameter file as input on command line
| Python | bsd-2-clause | jkitzes/macroeco |
+ import sys as _sys
from macroeco import desktop
- desktop()
+ if len(_sys.argv) > 1:
+ desktop(_sys.argv[1])
+ else:
+ desktop()
+ | Allow compiled OS X app to take parameter file as input on command line | ## Code Before:
from macroeco import desktop
desktop()
## Instruction:
Allow compiled OS X app to take parameter file as input on command line
## Code After:
import sys as _sys
from macroeco import desktop
if len(_sys.argv) > 1:
desktop(_sys.argv[1])
else:
desktop()
| ...
import sys as _sys
from macroeco import desktop
if len(_sys.argv) > 1:
desktop(_sys.argv[1])
else:
desktop()
... |
f3ea9820a96536e74e6f74f13387140c97ea9f2e | backgroundworker.py | backgroundworker.py | import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = engine, if_exists = 'replace')
| import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = connection, if_exists = 'replace')
connection.close()
| Add engine connection, and close engine connection to worker dyno | Add engine connection, and close engine connection to worker dyno
| Python | mit | caseymacphee/green_quote,caseymacphee/green_quote | import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
+
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
+ connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
- df.to_sql(name='entries', con = engine, if_exists = 'replace')
+ df.to_sql(name='entries', con = connection, if_exists = 'replace')
+ connection.close()
+ | Add engine connection, and close engine connection to worker dyno | ## Code Before:
import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = engine, if_exists = 'replace')
## Instruction:
Add engine connection, and close engine connection to worker dyno
## Code After:
import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = connection, if_exists = 'replace')
connection.close()
| // ... existing code ...
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = connection, if_exists = 'replace')
connection.close()
// ... rest of the code ... |
24fc06d17303868ef4ea057cd001ec6cb49ab18f | flask_app.py | flask_app.py | import os
from flask import Flask, render_template
from jinja2 import Template
app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..')
app.config.from_pyfile('settings.py')
BASE = '/%s' % app.config['REPO_NAME']
@app.route('/')
def home():
with open('talk.md', 'r') as f:
template = Template(f.read())
markdown = template.render(base=BASE)
js_file = 'talk.js'
if os.path.isfile(js_file):
with open(js_file, 'r') as f_js:
js = f_js.read()
else:
js = ''
return render_template('slides.html', markdown=markdown, js=js)
if __name__ == '__main__':
BASE = ''
port = int(os.environ.get('PORT', 5000))
app.run(host='0.0.0.0', port=port)
| import os
from flask import Flask, render_template
from jinja2 import Template
app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..')
app.config.from_pyfile('settings.py')
BASE = '/%s' % app.config['REPO_NAME']
@app.route('/')
def home():
with open('talk.md', 'r') as f:
template = Template(f.read().decode('utf-8'))
markdown = template.render(base=BASE)
js_file = 'talk.js'
if os.path.isfile(js_file):
with open(js_file, 'r') as f_js:
js = f_js.read()
else:
js = ''
return render_template('slides.html', markdown=markdown, js=js)
if __name__ == '__main__':
BASE = ''
port = int(os.environ.get('PORT', 5000))
app.run(host='0.0.0.0', port=port)
| Fix utf-8 problem with åäö and friends. | Fix utf-8 problem with åäö and friends.
| Python | bsd-3-clause | sknippen/refreeze,sknippen/refreeze,sknippen/refreeze | import os
from flask import Flask, render_template
from jinja2 import Template
app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..')
app.config.from_pyfile('settings.py')
BASE = '/%s' % app.config['REPO_NAME']
@app.route('/')
def home():
with open('talk.md', 'r') as f:
- template = Template(f.read())
+ template = Template(f.read().decode('utf-8'))
markdown = template.render(base=BASE)
js_file = 'talk.js'
if os.path.isfile(js_file):
with open(js_file, 'r') as f_js:
js = f_js.read()
else:
js = ''
return render_template('slides.html', markdown=markdown, js=js)
if __name__ == '__main__':
BASE = ''
port = int(os.environ.get('PORT', 5000))
app.run(host='0.0.0.0', port=port)
| Fix utf-8 problem with åäö and friends. | ## Code Before:
import os
from flask import Flask, render_template
from jinja2 import Template
app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..')
app.config.from_pyfile('settings.py')
BASE = '/%s' % app.config['REPO_NAME']
@app.route('/')
def home():
with open('talk.md', 'r') as f:
template = Template(f.read())
markdown = template.render(base=BASE)
js_file = 'talk.js'
if os.path.isfile(js_file):
with open(js_file, 'r') as f_js:
js = f_js.read()
else:
js = ''
return render_template('slides.html', markdown=markdown, js=js)
if __name__ == '__main__':
BASE = ''
port = int(os.environ.get('PORT', 5000))
app.run(host='0.0.0.0', port=port)
## Instruction:
Fix utf-8 problem with åäö and friends.
## Code After:
import os
from flask import Flask, render_template
from jinja2 import Template
app = Flask(__name__, template_folder='.', static_url_path='', static_folder='..')
app.config.from_pyfile('settings.py')
BASE = '/%s' % app.config['REPO_NAME']
@app.route('/')
def home():
with open('talk.md', 'r') as f:
template = Template(f.read().decode('utf-8'))
markdown = template.render(base=BASE)
js_file = 'talk.js'
if os.path.isfile(js_file):
with open(js_file, 'r') as f_js:
js = f_js.read()
else:
js = ''
return render_template('slides.html', markdown=markdown, js=js)
if __name__ == '__main__':
BASE = ''
port = int(os.environ.get('PORT', 5000))
app.run(host='0.0.0.0', port=port)
| // ... existing code ...
def home():
with open('talk.md', 'r') as f:
template = Template(f.read().decode('utf-8'))
markdown = template.render(base=BASE)
js_file = 'talk.js'
// ... rest of the code ... |
41b241de6f2afa94b442007518d481526bfb66ae | linked-list/remove-k-from-list.py | linked-list/remove-k-from-list.py |
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
|
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| Add initialization to linked list class | Add initialization to linked list class
| Python | mit | derekmpham/interview-prep,derekmpham/interview-prep |
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
+ class LinkedList(object):
+ def __init__(self, head=None):
+ self.head = head
+ | Add initialization to linked list class | ## Code Before:
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
## Instruction:
Add initialization to linked list class
## Code After:
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| # ... existing code ...
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
# ... rest of the code ... |
697c590bf60c261280e55f8580b33423dbe800c6 | splinter/driver/webdriver/firefox.py | splinter/driver/webdriver/firefox.py | import subprocess
from selenium.webdriver import Firefox
from selenium.webdriver.firefox.firefox_profile import FirefoxProfile
from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement
from splinter.driver.webdriver.cookie_manager import CookieManager
class WebDriver(BaseWebDriver):
def __init__(self, profile=None, extensions=None):
self.old_popen = subprocess.Popen
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
if extensions:
for extension in extensions:
firefox_profile.add_extension(extension)
self._patch_subprocess()
self.driver = Firefox(firefox_profile)
self._unpatch_subprocess()
self.element_class = WebDriverElement
self._cookie_manager = CookieManager(self.driver)
super(WebDriver, self).__init__()
class WebDriverElement(BaseWebDriverElement):
def mouseover(self):
"""
Firefox doesn't support mouseover.
"""
raise NotImplementedError("Firefox doesn't support mouse over")
def mouseout(self):
"""
Firefox doesn't support mouseout.
"""
raise NotImplementedError("Firefox doesn't support mouseout")
def double_click(self):
"""
Firefox doesn't support doubleclick.
"""
raise NotImplementedError("Firefox doesn't support doubleclick")
| import subprocess
from selenium.webdriver import Firefox
from selenium.webdriver.firefox.firefox_profile import FirefoxProfile
from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement
from splinter.driver.webdriver.cookie_manager import CookieManager
class WebDriver(BaseWebDriver):
def __init__(self, profile=None, extensions=None):
self.old_popen = subprocess.Popen
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
firefox_profile.set_preference('network.dns.disableIPv6', False)
if extensions:
for extension in extensions:
firefox_profile.add_extension(extension)
self._patch_subprocess()
self.driver = Firefox(firefox_profile)
self._unpatch_subprocess()
self.element_class = WebDriverElement
self._cookie_manager = CookieManager(self.driver)
super(WebDriver, self).__init__()
class WebDriverElement(BaseWebDriverElement):
def mouseover(self):
"""
Firefox doesn't support mouseover.
"""
raise NotImplementedError("Firefox doesn't support mouse over")
def mouseout(self):
"""
Firefox doesn't support mouseout.
"""
raise NotImplementedError("Firefox doesn't support mouseout")
def double_click(self):
"""
Firefox doesn't support doubleclick.
"""
raise NotImplementedError("Firefox doesn't support doubleclick")
| Fix error on Firefox 6 where pages are not open if this preference is True (default). | Fix error on Firefox 6 where pages are not open if this preference is True (default).
| Python | bsd-3-clause | bmcculley/splinter,cobrateam/splinter,bmcculley/splinter,nikolas/splinter,drptbl/splinter,objarni/splinter,nikolas/splinter,cobrateam/splinter,drptbl/splinter,underdogio/splinter,underdogio/splinter,bubenkoff/splinter,lrowe/splinter,bubenkoff/splinter,lrowe/splinter,objarni/splinter,gjvis/splinter,bmcculley/splinter,underdogio/splinter,objarni/splinter,lrowe/splinter,nikolas/splinter,cobrateam/splinter,gjvis/splinter,myself659/splinter,myself659/splinter,myself659/splinter,gjvis/splinter,drptbl/splinter | import subprocess
from selenium.webdriver import Firefox
from selenium.webdriver.firefox.firefox_profile import FirefoxProfile
from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement
from splinter.driver.webdriver.cookie_manager import CookieManager
class WebDriver(BaseWebDriver):
def __init__(self, profile=None, extensions=None):
self.old_popen = subprocess.Popen
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
+ firefox_profile.set_preference('network.dns.disableIPv6', False)
if extensions:
for extension in extensions:
firefox_profile.add_extension(extension)
self._patch_subprocess()
self.driver = Firefox(firefox_profile)
self._unpatch_subprocess()
self.element_class = WebDriverElement
self._cookie_manager = CookieManager(self.driver)
super(WebDriver, self).__init__()
class WebDriverElement(BaseWebDriverElement):
def mouseover(self):
"""
Firefox doesn't support mouseover.
"""
raise NotImplementedError("Firefox doesn't support mouse over")
def mouseout(self):
"""
Firefox doesn't support mouseout.
"""
raise NotImplementedError("Firefox doesn't support mouseout")
def double_click(self):
"""
Firefox doesn't support doubleclick.
"""
raise NotImplementedError("Firefox doesn't support doubleclick")
| Fix error on Firefox 6 where pages are not open if this preference is True (default). | ## Code Before:
import subprocess
from selenium.webdriver import Firefox
from selenium.webdriver.firefox.firefox_profile import FirefoxProfile
from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement
from splinter.driver.webdriver.cookie_manager import CookieManager
class WebDriver(BaseWebDriver):
def __init__(self, profile=None, extensions=None):
self.old_popen = subprocess.Popen
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
if extensions:
for extension in extensions:
firefox_profile.add_extension(extension)
self._patch_subprocess()
self.driver = Firefox(firefox_profile)
self._unpatch_subprocess()
self.element_class = WebDriverElement
self._cookie_manager = CookieManager(self.driver)
super(WebDriver, self).__init__()
class WebDriverElement(BaseWebDriverElement):
def mouseover(self):
"""
Firefox doesn't support mouseover.
"""
raise NotImplementedError("Firefox doesn't support mouse over")
def mouseout(self):
"""
Firefox doesn't support mouseout.
"""
raise NotImplementedError("Firefox doesn't support mouseout")
def double_click(self):
"""
Firefox doesn't support doubleclick.
"""
raise NotImplementedError("Firefox doesn't support doubleclick")
## Instruction:
Fix error on Firefox 6 where pages are not open if this preference is True (default).
## Code After:
import subprocess
from selenium.webdriver import Firefox
from selenium.webdriver.firefox.firefox_profile import FirefoxProfile
from splinter.driver.webdriver import BaseWebDriver, WebDriverElement as BaseWebDriverElement
from splinter.driver.webdriver.cookie_manager import CookieManager
class WebDriver(BaseWebDriver):
def __init__(self, profile=None, extensions=None):
self.old_popen = subprocess.Popen
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
firefox_profile.set_preference('network.dns.disableIPv6', False)
if extensions:
for extension in extensions:
firefox_profile.add_extension(extension)
self._patch_subprocess()
self.driver = Firefox(firefox_profile)
self._unpatch_subprocess()
self.element_class = WebDriverElement
self._cookie_manager = CookieManager(self.driver)
super(WebDriver, self).__init__()
class WebDriverElement(BaseWebDriverElement):
def mouseover(self):
"""
Firefox doesn't support mouseover.
"""
raise NotImplementedError("Firefox doesn't support mouse over")
def mouseout(self):
"""
Firefox doesn't support mouseout.
"""
raise NotImplementedError("Firefox doesn't support mouseout")
def double_click(self):
"""
Firefox doesn't support doubleclick.
"""
raise NotImplementedError("Firefox doesn't support doubleclick")
| # ... existing code ...
firefox_profile = FirefoxProfile(profile)
firefox_profile.set_preference('extensions.logging.enabled', False)
firefox_profile.set_preference('network.dns.disableIPv6', False)
if extensions:
# ... rest of the code ... |
ea8cbcaf41f01a46390882fbc99e6e14d70a49d1 | src/mmw/apps/user/models.py | src/mmw/apps/user/models.py | from django.contrib.auth.models import User
from django.db import models
class ItsiUserManager(models.Manager):
def create_itsi_user(self, user, itsi_id):
itsi_user = self.create(user=user, itsi_id=itsi_id)
return itsi_user
class ItsiUser(models.Model):
user = models.OneToOneField(User, primary_key=True)
itsi_id = models.IntegerField()
objects = ItsiUserManager()
def __unicode__(self):
return unicode(self.user.username)
|
from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
from django.db.models.signals import post_save
from django.dispatch import receiver
from rest_framework.authtoken.models import Token
@receiver(post_save, sender=settings.AUTH_USER_MODEL)
def create_auth_token(sender, instance=None, created=False, **kwargs):
"""
Create an auth token for every newly created user.
"""
if created:
Token.objects.create(user=instance)
class ItsiUserManager(models.Manager):
def create_itsi_user(self, user, itsi_id):
itsi_user = self.create(user=user, itsi_id=itsi_id)
return itsi_user
class ItsiUser(models.Model):
user = models.OneToOneField(User, primary_key=True)
itsi_id = models.IntegerField()
objects = ItsiUserManager()
def __unicode__(self):
return unicode(self.user.username)
| Create an API auth token for every newly created user | Create an API auth token for every newly created user
* Add a post_save signal to add a new authtoken for every new user. For use with
the Geoprocessing API
| Python | apache-2.0 | WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed | +
+ from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
+ from django.db.models.signals import post_save
+ from django.dispatch import receiver
+
+ from rest_framework.authtoken.models import Token
+
+
+ @receiver(post_save, sender=settings.AUTH_USER_MODEL)
+ def create_auth_token(sender, instance=None, created=False, **kwargs):
+ """
+ Create an auth token for every newly created user.
+ """
+ if created:
+ Token.objects.create(user=instance)
class ItsiUserManager(models.Manager):
def create_itsi_user(self, user, itsi_id):
itsi_user = self.create(user=user, itsi_id=itsi_id)
return itsi_user
class ItsiUser(models.Model):
user = models.OneToOneField(User, primary_key=True)
itsi_id = models.IntegerField()
objects = ItsiUserManager()
def __unicode__(self):
return unicode(self.user.username)
| Create an API auth token for every newly created user | ## Code Before:
from django.contrib.auth.models import User
from django.db import models
class ItsiUserManager(models.Manager):
def create_itsi_user(self, user, itsi_id):
itsi_user = self.create(user=user, itsi_id=itsi_id)
return itsi_user
class ItsiUser(models.Model):
user = models.OneToOneField(User, primary_key=True)
itsi_id = models.IntegerField()
objects = ItsiUserManager()
def __unicode__(self):
return unicode(self.user.username)
## Instruction:
Create an API auth token for every newly created user
## Code After:
from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
from django.db.models.signals import post_save
from django.dispatch import receiver
from rest_framework.authtoken.models import Token
@receiver(post_save, sender=settings.AUTH_USER_MODEL)
def create_auth_token(sender, instance=None, created=False, **kwargs):
"""
Create an auth token for every newly created user.
"""
if created:
Token.objects.create(user=instance)
class ItsiUserManager(models.Manager):
def create_itsi_user(self, user, itsi_id):
itsi_user = self.create(user=user, itsi_id=itsi_id)
return itsi_user
class ItsiUser(models.Model):
user = models.OneToOneField(User, primary_key=True)
itsi_id = models.IntegerField()
objects = ItsiUserManager()
def __unicode__(self):
return unicode(self.user.username)
| # ... existing code ...
from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
from django.db.models.signals import post_save
from django.dispatch import receiver
from rest_framework.authtoken.models import Token
@receiver(post_save, sender=settings.AUTH_USER_MODEL)
def create_auth_token(sender, instance=None, created=False, **kwargs):
"""
Create an auth token for every newly created user.
"""
if created:
Token.objects.create(user=instance)
# ... rest of the code ... |
424980a48e451d1b99397843001bd75fa58e474e | tests/test_fullqualname.py | tests/test_fullqualname.py | """Tests for fullqualname."""
import nose
import sys
from fullqualname import fullqualname
def test_builtin_function():
# Test built-in function object.
obj = len
# Type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a function.
assert 'built-in function' in repr(obj)
if sys.version_info >= (3, ):
expected = 'builtins.len'
else:
expected = '__builtin__.len'
nose.tools.assert_equals(fullqualname(obj), expected)
| """Tests for fullqualname."""
import inspect
import nose
import sys
from fullqualname import fullqualname
def test_builtin_function():
# Test built-in function object.
obj = len
# Type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a function.
assert 'built-in function' in repr(obj)
if sys.version_info >= (3, ):
expected = 'builtins.len'
else:
expected = '__builtin__.len'
nose.tools.assert_equals(fullqualname(obj), expected)
def test_builtin_method():
# Test built-in method object.
obj = [1, 2, 3].append
# Object type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a method.
assert 'built-in method' in repr(obj)
# Object __self__ attribute is not a class.
assert not inspect.isclass(obj.__self__)
if sys.version_info >= (3, ):
expected = 'builtins.list.append'
else:
expected = '__builtin__.list.append'
nose.tools.assert_equals(fullqualname(obj), expected)
| Add built-in method object test | Add built-in method object test
| Python | bsd-3-clause | etgalloway/fullqualname | """Tests for fullqualname."""
+ import inspect
import nose
import sys
from fullqualname import fullqualname
def test_builtin_function():
# Test built-in function object.
obj = len
# Type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a function.
assert 'built-in function' in repr(obj)
if sys.version_info >= (3, ):
expected = 'builtins.len'
else:
expected = '__builtin__.len'
nose.tools.assert_equals(fullqualname(obj), expected)
+
+ def test_builtin_method():
+ # Test built-in method object.
+
+ obj = [1, 2, 3].append
+
+ # Object type is 'builtin_function_or_method'.
+ assert type(obj).__name__ == 'builtin_function_or_method'
+
+ # Object is a method.
+ assert 'built-in method' in repr(obj)
+
+ # Object __self__ attribute is not a class.
+ assert not inspect.isclass(obj.__self__)
+
+ if sys.version_info >= (3, ):
+ expected = 'builtins.list.append'
+ else:
+ expected = '__builtin__.list.append'
+
+ nose.tools.assert_equals(fullqualname(obj), expected)
+ | Add built-in method object test | ## Code Before:
"""Tests for fullqualname."""
import nose
import sys
from fullqualname import fullqualname
def test_builtin_function():
# Test built-in function object.
obj = len
# Type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a function.
assert 'built-in function' in repr(obj)
if sys.version_info >= (3, ):
expected = 'builtins.len'
else:
expected = '__builtin__.len'
nose.tools.assert_equals(fullqualname(obj), expected)
## Instruction:
Add built-in method object test
## Code After:
"""Tests for fullqualname."""
import inspect
import nose
import sys
from fullqualname import fullqualname
def test_builtin_function():
# Test built-in function object.
obj = len
# Type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a function.
assert 'built-in function' in repr(obj)
if sys.version_info >= (3, ):
expected = 'builtins.len'
else:
expected = '__builtin__.len'
nose.tools.assert_equals(fullqualname(obj), expected)
def test_builtin_method():
# Test built-in method object.
obj = [1, 2, 3].append
# Object type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a method.
assert 'built-in method' in repr(obj)
# Object __self__ attribute is not a class.
assert not inspect.isclass(obj.__self__)
if sys.version_info >= (3, ):
expected = 'builtins.list.append'
else:
expected = '__builtin__.list.append'
nose.tools.assert_equals(fullqualname(obj), expected)
| # ... existing code ...
"""Tests for fullqualname."""
import inspect
import nose
import sys
# ... modified code ...
nose.tools.assert_equals(fullqualname(obj), expected)
def test_builtin_method():
# Test built-in method object.
obj = [1, 2, 3].append
# Object type is 'builtin_function_or_method'.
assert type(obj).__name__ == 'builtin_function_or_method'
# Object is a method.
assert 'built-in method' in repr(obj)
# Object __self__ attribute is not a class.
assert not inspect.isclass(obj.__self__)
if sys.version_info >= (3, ):
expected = 'builtins.list.append'
else:
expected = '__builtin__.list.append'
nose.tools.assert_equals(fullqualname(obj), expected)
# ... rest of the code ... |
85e8ddb6d72b7f21b49236ea4084029dec09a6f9 | projects/forms.py | projects/forms.py | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
fileds = (
'status',
'attitude', ) | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| Exclude fields from the RestrcitedForm (no verification) | Exclude fields from the RestrcitedForm (no verification)
| Python | mit | Hackfmi/Diaphanum,Hackfmi/Diaphanum | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
+ exclude = (
+ 'name', 'team', 'description', 'targets', 'tasks', 'target_group',
+ 'schedule', 'resources', 'finance_description', 'partners',
+ 'flp', 'created_at', 'user',
+ )
+
fileds = (
'status',
'attitude', )
+ | Exclude fields from the RestrcitedForm (no verification) | ## Code Before:
from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
fileds = (
'status',
'attitude', )
## Instruction:
Exclude fields from the RestrcitedForm (no verification)
## Code After:
from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| # ... existing code ...
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
# ... rest of the code ... |
Subsets and Splits