iterative/dvc

View on GitHub

Showing 148 of 261 total issues

Similar blocks of code found in 6 locations. Consider refactoring.
Open

class DependencyDoesNotExistError(DvcException):
    def __init__(self, path):
        msg = "dependency '{}' does not exist".format(path)
        super(DependencyDoesNotExistError, self).__init__(msg)
Severity: Major
Found in dvc/dependency/base.py and 5 other locations - About 50 mins to fix
dvc/dependency/base.py on lines 12..15
dvc/output/base.py on lines 20..23
dvc/output/base.py on lines 26..29
dvc/output/base.py on lines 32..35
dvc/stage.py on lines 55..58

Duplicated Code

Duplicated code can lead to software that is hard to understand and difficult to change. The Don't Repeat Yourself (DRY) principle states:

Every piece of knowledge must have a single, unambiguous, authoritative representation within a system.

When you violate DRY, bugs and maintenance problems are sure to follow. Duplicated code has a tendency to both continue to replicate and also to diverge (leaving bugs as two similar implementations differ in subtle ways).

Tuning

This issue has a mass of 36.

We set useful threshold defaults for the languages we support but you may want to adjust these settings based on your project guidelines.

The threshold configuration represents the minimum mass a code block must have to be analyzed for duplication. The lower the threshold, the more fine-grained the comparison.

If the engine is too easily reporting duplication, try raising the threshold. If you suspect that the engine isn't catching enough duplication, try lowering the threshold. The best setting tends to differ from language to language.

See codeclimate-duplication's documentation for more information about tuning the mass threshold in your .codeclimate.yml.

Refactorings

Further Reading

Avoid deeply nested control flow statements.
Open

                    if not data:
                        break

Severity: Major
Found in dvc/utils/__init__.py - About 45 mins to fix

    Function _create_stages has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
    Open

    def _create_stages(repo, targets, fname, pbar=None):
        stages = []
    
        for out in targets:
            stage = Stage.create(repo, outs=[out], add=True, fname=fname)
    Severity: Minor
    Found in dvc/repo/add.py - About 45 mins to fix

    Cognitive Complexity

    Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

    A method's cognitive complexity is based on a few simple rules:

    • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
    • Code is considered more complex for each "break in the linear flow of the code"
    • Code is considered more complex when "flow breaking structures are nested"

    Further reading

    Function _copy_multipart has 6 arguments (exceeds 4 allowed). Consider refactoring.
    Open

        def _copy_multipart(
    Severity: Minor
    Found in dvc/remote/s3.py - About 45 mins to fix

      Function _show_dependencies_tree has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
      Open

          def _show_dependencies_tree(self, target, commands, outs):
              from treelib import Tree
      
              nodes, edges, is_tree = self.__build_graph(target, commands, outs)
              if not nodes:
      Severity: Minor
      Found in dvc/command/pipeline.py - About 45 mins to fix

      Cognitive Complexity

      Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

      A method's cognitive complexity is based on a few simple rules:

      • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
      • Code is considered more complex for each "break in the linear flow of the code"
      • Code is considered more complex when "flow breaking structures are nested"

      Further reading

      Function from_parts has 6 arguments (exceeds 4 allowed). Consider refactoring.
      Open

          def from_parts(
      Severity: Minor
      Found in dvc/path_info.py - About 45 mins to fix

        Function reproduce has 6 arguments (exceeds 4 allowed). Consider refactoring.
        Open

        def reproduce(
        Severity: Minor
        Found in dvc/repo/reproduce.py - About 45 mins to fix

          Function list_cache_paths has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
          Open

              def list_cache_paths(self):
                  if not self.exists(self.path_info):
                      return
          
                  dirs = deque([self.path_info.path])
          Severity: Minor
          Found in dvc/remote/hdfs.py - About 45 mins to fix

          Cognitive Complexity

          Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

          A method's cognitive complexity is based on a few simple rules:

          • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
          • Code is considered more complex for each "break in the linear flow of the code"
          • Code is considered more complex when "flow breaking structures are nested"

          Further reading

          Function convert_to_unicode has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
          Open

              def convert_to_unicode(data):
                  if isinstance(data, builtin_str):
                      return str(data)
                  if isinstance(data, dict):
                      return dict(map(convert_to_unicode, data.items()))
          Severity: Minor
          Found in dvc/utils/compat.py - About 45 mins to fix

          Cognitive Complexity

          Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

          A method's cognitive complexity is based on a few simple rules:

          • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
          • Code is considered more complex for each "break in the linear flow of the code"
          • Code is considered more complex when "flow breaking structures are nested"

          Further reading

          Function _upload_to_bucket has 6 arguments (exceeds 4 allowed). Consider refactoring.
          Open

          def _upload_to_bucket(
          Severity: Minor
          Found in dvc/remote/gs.py - About 45 mins to fix

            Function checkout has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
            Open

                def checkout(
                    self, path_info, checksum_info, force=False, progress_callback=None
                ):
                    if path_info.scheme not in ["local", self.scheme]:
                        raise NotImplementedError
            Severity: Minor
            Found in dvc/remote/base.py - About 45 mins to fix

            Cognitive Complexity

            Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

            A method's cognitive complexity is based on a few simple rules:

            • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
            • Code is considered more complex for each "break in the linear flow of the code"
            • Code is considered more complex when "flow breaking structures are nested"

            Further reading

            Function open has 6 arguments (exceeds 4 allowed). Consider refactoring.
            Open

            def open(path, repo=None, rev=None, remote=None, mode="r", encoding=None):
            Severity: Minor
            Found in dvc/api.py - About 45 mins to fix

              Function _collect_metrics has 6 arguments (exceeds 4 allowed). Consider refactoring.
              Open

              def _collect_metrics(repo, path, recursive, typ, xpath, branch):
              Severity: Minor
              Found in dvc/repo/metrics/show.py - About 45 mins to fix

                Avoid deeply nested control flow statements.
                Open

                                    if binary:
                                        chunk = data
                                    else:
                                        chunk = dos2unix(data)
                
                
                Severity: Major
                Found in dvc/utils/__init__.py - About 45 mins to fix

                  Function draw has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
                  Open

                  def draw(vertexes, edges):
                      """Build a DAG and draw it in ASCII.
                  
                      Args:
                          vertexes (list): list of graph vertexes.
                  Severity: Minor
                  Found in dvc/dagascii.py - About 45 mins to fix

                  Cognitive Complexity

                  Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

                  A method's cognitive complexity is based on a few simple rules:

                  • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
                  • Code is considered more complex for each "break in the linear flow of the code"
                  • Code is considered more complex when "flow breaking structures are nested"

                  Further reading

                  Identical blocks of code found in 2 locations. Consider refactoring.
                  Open

                          with Tqdm(
                              desc="Querying "
                              + ("cache in " + name if name else "remote cache"),
                              total=len(checksums),
                  Severity: Minor
                  Found in dvc/remote/base.py and 1 other location - About 45 mins to fix
                  dvc/remote/ssh/__init__.py on lines 318..321

                  Duplicated Code

                  Duplicated code can lead to software that is hard to understand and difficult to change. The Don't Repeat Yourself (DRY) principle states:

                  Every piece of knowledge must have a single, unambiguous, authoritative representation within a system.

                  When you violate DRY, bugs and maintenance problems are sure to follow. Duplicated code has a tendency to both continue to replicate and also to diverge (leaving bugs as two similar implementations differ in subtle ways).

                  Tuning

                  This issue has a mass of 35.

                  We set useful threshold defaults for the languages we support but you may want to adjust these settings based on your project guidelines.

                  The threshold configuration represents the minimum mass a code block must have to be analyzed for duplication. The lower the threshold, the more fine-grained the comparison.

                  If the engine is too easily reporting duplication, try raising the threshold. If you suspect that the engine isn't catching enough duplication, try lowering the threshold. The best setting tends to differ from language to language.

                  See codeclimate-duplication's documentation for more information about tuning the mass threshold in your .codeclimate.yml.

                  Refactorings

                  Further Reading

                  Similar blocks of code found in 3 locations. Consider refactoring.
                  Open

                      push_parser = subparsers.add_parser(
                          "push",
                          parents=[shared_parent_parser()],
                          description=append_doc_link(PUSH_HELP, "push"),
                          help=PUSH_HELP,
                  Severity: Major
                  Found in dvc/command/data_sync.py and 2 other locations - About 45 mins to fix
                  dvc/command/data_sync.py on lines 105..110
                  dvc/command/data_sync.py on lines 198..203

                  Duplicated Code

                  Duplicated code can lead to software that is hard to understand and difficult to change. The Don't Repeat Yourself (DRY) principle states:

                  Every piece of knowledge must have a single, unambiguous, authoritative representation within a system.

                  When you violate DRY, bugs and maintenance problems are sure to follow. Duplicated code has a tendency to both continue to replicate and also to diverge (leaving bugs as two similar implementations differ in subtle ways).

                  Tuning

                  This issue has a mass of 35.

                  We set useful threshold defaults for the languages we support but you may want to adjust these settings based on your project guidelines.

                  The threshold configuration represents the minimum mass a code block must have to be analyzed for duplication. The lower the threshold, the more fine-grained the comparison.

                  If the engine is too easily reporting duplication, try raising the threshold. If you suspect that the engine isn't catching enough duplication, try lowering the threshold. The best setting tends to differ from language to language.

                  See codeclimate-duplication's documentation for more information about tuning the mass threshold in your .codeclimate.yml.

                  Refactorings

                  Further Reading

                  Identical blocks of code found in 2 locations. Consider refactoring.
                  Open

                          if not self.no_traverse:
                              return list(set(checksums) & set(self.all()))
                  Severity: Minor
                  Found in dvc/remote/ssh/__init__.py and 1 other location - About 45 mins to fix
                  dvc/remote/base.py on lines 783..784

                  Duplicated Code

                  Duplicated code can lead to software that is hard to understand and difficult to change. The Don't Repeat Yourself (DRY) principle states:

                  Every piece of knowledge must have a single, unambiguous, authoritative representation within a system.

                  When you violate DRY, bugs and maintenance problems are sure to follow. Duplicated code has a tendency to both continue to replicate and also to diverge (leaving bugs as two similar implementations differ in subtle ways).

                  Tuning

                  This issue has a mass of 35.

                  We set useful threshold defaults for the languages we support but you may want to adjust these settings based on your project guidelines.

                  The threshold configuration represents the minimum mass a code block must have to be analyzed for duplication. The lower the threshold, the more fine-grained the comparison.

                  If the engine is too easily reporting duplication, try raising the threshold. If you suspect that the engine isn't catching enough duplication, try lowering the threshold. The best setting tends to differ from language to language.

                  See codeclimate-duplication's documentation for more information about tuning the mass threshold in your .codeclimate.yml.

                  Refactorings

                  Further Reading

                  Identical blocks of code found in 2 locations. Consider refactoring.
                  Open

                          with Tqdm(
                              desc="Querying "
                              + ("cache in " + name if name else "remote cache"),
                              total=len(checksums),
                  Severity: Minor
                  Found in dvc/remote/ssh/__init__.py and 1 other location - About 45 mins to fix
                  dvc/remote/base.py on lines 786..789

                  Duplicated Code

                  Duplicated code can lead to software that is hard to understand and difficult to change. The Don't Repeat Yourself (DRY) principle states:

                  Every piece of knowledge must have a single, unambiguous, authoritative representation within a system.

                  When you violate DRY, bugs and maintenance problems are sure to follow. Duplicated code has a tendency to both continue to replicate and also to diverge (leaving bugs as two similar implementations differ in subtle ways).

                  Tuning

                  This issue has a mass of 35.

                  We set useful threshold defaults for the languages we support but you may want to adjust these settings based on your project guidelines.

                  The threshold configuration represents the minimum mass a code block must have to be analyzed for duplication. The lower the threshold, the more fine-grained the comparison.

                  If the engine is too easily reporting duplication, try raising the threshold. If you suspect that the engine isn't catching enough duplication, try lowering the threshold. The best setting tends to differ from language to language.

                  See codeclimate-duplication's documentation for more information about tuning the mass threshold in your .codeclimate.yml.

                  Refactorings

                  Further Reading

                  Function update has a Cognitive Complexity of 8 (exceeds 5 allowed). Consider refactoring.
                  Open

                      def update(self, cache, suffix=""):
                          for scheme, src in cache._items.items():
                              dst = self._items[scheme]
                              for checksum, names in src.items():
                                  if suffix:
                  Severity: Minor
                  Found in dvc/cache.py - About 45 mins to fix

                  Cognitive Complexity

                  Cognitive Complexity is a measure of how difficult a unit of code is to intuitively understand. Unlike Cyclomatic Complexity, which determines how difficult your code will be to test, Cognitive Complexity tells you how difficult your code will be to read and comprehend.

                  A method's cognitive complexity is based on a few simple rules:

                  • Code is not considered more complex when it uses shorthand that the language provides for collapsing multiple statements into one
                  • Code is considered more complex for each "break in the linear flow of the code"
                  • Code is considered more complex when "flow breaking structures are nested"

                  Further reading

                  Severity
                  Category
                  Status
                  Source
                  Language