Browse Source

unit-test exclusion

master
lhenry15 4 years ago
parent
commit
68c0e68c45
3 changed files with 12 additions and 12 deletions
  1. +1
    -1
      tods/schemas.py
  2. +7
    -7
      tods/searcher/brute_force_search.py
  3. +4
    -4
      tods/utils.py

+ 1
- 1
tods/schemas.py View File

@@ -4,7 +4,7 @@ resource_dir = os.path.dirname(__file__)

DEFAULT_PIPELINE_DIR = os.path.join(resource_dir, 'resources', 'default_pipeline.json')

def load_default_pipeline():
def load_default_pipeline(): # pragma: no cover
from axolotl.utils import pipeline as pipeline_utils
pipeline = pipeline_utils.load_pipeline(DEFAULT_PIPELINE_DIR)
return pipeline

+ 7
- 7
tods/searcher/brute_force_search.py View File

@@ -7,7 +7,7 @@ from d3m.metadata.pipeline import Pipeline
from axolotl.algorithms.base import PipelineSearchBase
from axolotl.utils import schemas as schemas_utils

class BruteForceSearch(PipelineSearchBase):
class BruteForceSearch(PipelineSearchBase): # pragma: no cover
def __init__(self, problem_description, backend, *, primitives_blocklist=None, ranking_function=None):
super().__init__(problem_description=problem_description, backend=backend,
primitives_blocklist=primitives_blocklist, ranking_function=ranking_function)
@@ -153,7 +153,7 @@ primitive_python_paths = {
}


def _rank_first_metric(pipeline_result):
def _rank_first_metric(pipeline_result): # pragma: no cover
if pipeline_result.status == 'COMPLETED':
scores = pipeline_result.scores
pipeline_result.rank = -scores['value'][0]
@@ -163,22 +163,22 @@ def _rank_first_metric(pipeline_result):
pipeline_result.rank = 1
return pipeline_result

def _generate_data_preparation_params():
def _generate_data_preparation_params(): # pragma: no cover
from axolotl.utils import schemas as schemas_utils
data_preparation_params = schemas_utils.DATA_PREPARATION_PARAMS['no_split']
return data_preparation_params
def _generate_scoring_pipeline():
def _generate_scoring_pipeline(): # pragma: no cover
from axolotl.utils import schemas as schemas_utils
scoring_pipeline = schemas_utils.get_scoring_pipeline()
return scoring_pipeline
def _generate_data_preparation_pipeline():
def _generate_data_preparation_pipeline(): # pragma: no cover
from axolotl.utils import schemas as schemas_utils
data_preparation_pipeline = schemas_utils.get_splitting_pipeline("TRAINING_DATA")
return data_preparation_pipeline

def _generate_pipline(combinations):
def _generate_pipline(combinations): # pragma: no cover
from d3m import index
from d3m.metadata.base import ArgumentType
from d3m.metadata.pipeline import Pipeline, PrimitiveStep
@@ -257,7 +257,7 @@ def _generate_pipline(combinations):
piplines.append(pipeline_description)
return piplines

def _generate_pipelines(primitive_python_paths, cpu_count=40):
def _generate_pipelines(primitive_python_paths, cpu_count=40): # pragma: no cover
"""
Args:
primitive_python_paths: a list of primitive Python paths for algorithms


+ 4
- 4
tods/utils.py View File

@@ -1,5 +1,5 @@

def load_pipeline(pipeline_path):
def load_pipeline(pipeline_path): # pragma: no cover
"""Load a pipeline given a path

Args:
@@ -13,7 +13,7 @@ def load_pipeline(pipeline_path):

return pipeline
def generate_dataset(df, target_index, system_dir=None):
def generate_dataset(df, target_index, system_dir=None): # pragma: no cover
"""Generate dataset

Args:
@@ -29,7 +29,7 @@ def generate_dataset(df, target_index, system_dir=None):

return dataset

def generate_problem(dataset, metric):
def generate_problem(dataset, metric): # pragma: no cover
"""Generate dataset

Args:
@@ -61,7 +61,7 @@ def generate_problem(dataset, metric):
return problem_description

def evaluate_pipeline(dataset, pipeline, metric='F1', seed=0):
def evaluate_pipeline(dataset, pipeline, metric='F1', seed=0): # pragma: no cover
"""Evaluate a Pipeline

Args:


Loading…
Cancel
Save