/external/autotest/venv/lucifer/ |
jobx.py | 33 return [hqe.host.hostname for hqe in hqes if hqe.host is not None] 44 for hqe in job.hostqueueentry_set.all(): 45 if hqe.aborted:
|
handlers.py | 98 for hqe in self._job.hostqueueentry_set.all(): 99 self._set_completed_status(hqe, final_status) 155 def _set_completed_status(self, hqe, status): 156 """Set completed status of HQE. 161 hqe.status = status 162 hqe.active = False 163 hqe.complete = True 164 if hqe.started_on: 165 hqe.finished_on = datetime.datetime.now() 166 hqe.save( [all...] |
/external/autotest/scheduler/ |
scheduler_models_unittest.py | 116 hqe = scheduler_models.HostQueueEntry( 120 hqe.save() 121 new_id = hqe.id 124 hqe = scheduler_models.HostQueueEntry(id=new_id) 125 self.assertEqual(hqe.id, new_id) 126 self.assertEqual(hqe.job_id, 1) 127 self.assertEqual(hqe.host_id, 2) 128 self.assertEqual(hqe.status, 'Queued') 129 self.assertEqual(hqe.meta_host, None) 130 self.assertEqual(hqe.active, False [all...] |
host_scheduler_unittests.py | 82 # HQE becomes active and the host remains assigned to it. 101 # HQE becomes DOES NOT become active, because we validate the 121 # HQE does not become active through the scheduler, and that the 146 # Create 2 special tasks, one with and one without an hqe. 147 # Activate the hqe and make sure it gets scheduled before the other. 152 hqe = self.db_helper.get_hqes(job=job1.id)[0] 157 # Since the hqe task isn't active we get both back. 160 tasks[0].queue_entry_id == hqe.id) 162 # Activate the hqe and make sure the frontned task isn't returned. 163 self.db_helper.update_hqe(hqe.id, active=True [all...] |
monitor_db_unittest.py | 394 def fake_set_status(hqe, *args, **kwargs): 395 self.assertEqual(hqe.execution_subdir, 'hostless') 396 original_set_status(hqe, *args, **kwargs) 403 hqe = job.hostqueueentry_set.all()[0] 404 self.assertEqual(models.HostQueueEntry.Status.STARTING, hqe.status) 405 self.assertEqual('hostless', hqe.execution_subdir) 797 hqe_ids = [hqe.id for hqe in queue_task.queue_entries] 831 hqe = hqe_query[0] 833 self.assertEqual(models.HostQueueEntry.Status.QUEUED, hqe.status [all...] |
rdb_testing_utils.py | 240 """Update the hqe with the given kwargs. 242 @param hqe_id: The id of the hqe to update. 258 """Add a host to the hqe of a job. 262 @param activate: If true, flip the active bit on the hqe. 264 @raises ValueError: If the hqe for the job already has a host, 267 hqe = models.HostQueueEntry.objects.get(job_id=job_id) 268 if hqe.host: 269 raise ValueError('HQE for job %s already has a host' % job_id) 270 hqe.host = host 271 hqe.save( [all...] |
query_managers.py | 49 activate the hqe after assigning a host to it in 161 @return: A list of dictionaries with the hqe id, job_id and host_id 166 # stand alone special task can share a host with an active hqe, an 198 for hqe in query: 199 host_id = hqe.host_id 200 parent_job_id = hqe.job.parent_job_id 315 # this tick: 1. When the host is locked 2. When an active hqe already
|
luciferlib.py | 53 """Return True if HQE is part of a job with HQEs in a different group. 55 For examples if the given HQE have execution_subdir=foo and the job 56 has an HQE with execution_subdir=bar, then return True. The only 62 @param hqe_id: HQE id 64 hqe = models.HostQueueEntry.objects.get(id=hqe_id) 65 hqes = hqe.job.hostqueueentry_set.all()
|
rdb_integration_tests.py | 189 been assigned to an active hqe. 197 # Create a job and give its hqe the leased host. 200 hqe = models.HostQueueEntry.objects.get(job_id=job.id) 202 # Activate the hqe by setting its state. 203 hqe.status = host_queue_entry_states.ACTIVE_STATUSES[0] 204 hqe.save()
|
monitor_db.py | 490 # TODO(crbug.com/811877): Don't skip split HQE parsing. 499 # TODO(crbug.com/811877): Don't skip split HQE parsing. 569 Get agent tasks for all hqe in the specified states. 571 Loosely this translates to taking a hqe in one of the specified states, 611 logging.exception('Skipping agent task for a malformed hqe.') 613 # this errant HQE. It appears that calling entry.abort() is not 696 and a Null hqe. To create a SpecialAgentTask object, you need a 698 object contains a hqe it's passed on to the special agent task, which [all...] |
/external/autotest/site_utils/ |
job_history.py | 188 def __init__(self, hqe, task_caches=None, suite_start_time=None, 192 @param hqe: HostQueueEntry of the job. 200 self.id = hqe.job.id 202 self.job = hqe.job 204 self.name = hqe.job.name.split('/')[-1] 205 self.status = hqe.status if hqe else None 228 self.log_url = JOB_URL % {'job_id': hqe.job.id, 'owner': hqe.job.owner, 232 self._get_special_tasks(hqe, task_caches, suite_start_time [all...] |
lxc_cleanup.py | 67 logging.error('Failed to get hqe for job %s. Error: %s.', job_id, e) 73 for hqe in hqes: 74 if hqe.active or not hqe.complete: 78 if (hqe.finished_on and 79 (time_utils.time_string_to_datetime(hqe.finished_on) >
|
job_directories.py | 190 @returns the latest hqe finished_on time. If the finished_on times are null 200 # While most Jobs have 1 HQE, some can have multiple, so check them all. 201 return max([hqe.finished_on for hqe in hqes])
|
test_push.py | 389 hostnames = set([hqe.host.hostname for hqe in hqes]) 414 # hqe.host_id for jobs running in shard is not synced back to master db,
|
/external/autotest/server/lib/ |
suite_report.py | 34 """Parse a status string from TKO or the HQE databases. 36 @param status_string: A status string from TKO or HQE databases. 134 def make_hqe_entry(hostname, hqe, hqe_statuses, parent=None): 135 """Generate a HQE event log entry. 138 @param hqe: A host history to generate an event for. 139 @param hqe_statuses: A dictionary mapping HQE ids to job status. 145 ('HQE', int(hqe.id)), hostname, 146 hqe_statuses.get(hqe.id, parse_tko_status_string(hqe.job_status)) [all...] |
status_history.py | 387 return [cls(afe.server, hqe) for hqe in hqelist] 390 def __init__(self, afe_hostname, hqe): 392 self._hqe = hqe 394 hqe.started_on, hqe.finished_on)
|
/external/autotest/scheduler/shard/ |
shard_client_integration_tests.py | 44 hqe = scheduler_models.HostQueueEntry.fetch( 48 hqe.set_status('Completed') 66 hqe = scheduler_models.HostQueueEntry.fetch( 69 def _local_update_field(hqe, field_name, value): 73 models.HostQueueEntry.objects.filter(id=hqe.id).update( 75 setattr(hqe, field_name, value) 81 hqe.set_status('Completed') 83 # Retrieve the hqe along an independent code path so we're assured of 86 assert(modified_hqe.id == hqe.id and
|
shard_client_unittest.py | 183 job, hqe = jobs[0], hqes[0] 184 self.assertEqual(hqe['status'], 'Completed') 215 hqe = job.hostqueueentry_set.all()[0] 216 hqe.status = 'Completed' 217 hqe.save()
|
shard_client.py | 301 hqes = [hqe.serialize(include_dependencies=False) 302 for hqe in self._get_hqes_for_jobs(job_objs)]
|
/external/autotest/server/hosts/ |
servo_host.py | 392 for hqe in hqes: 393 job = afe.get_jobs(id=hqe.job.id) [all...] |
/external/autotest/frontend/afe/ |
rpc_interface_unittest.py | 80 expected_hqes = [(hqe.id) for hqe in hqes] 81 returned_hqes = [(hqe['id']) for hqe in retval_hqes] [all...] |
rpc_interface.py | [all...] |