Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat: add pairtab compression #4432

Merged
merged 4 commits into from
Nov 27, 2024

Conversation

anyangml
Copy link
Collaborator

@anyangml anyangml commented Nov 27, 2024

Summary by CodeRabbit

  • New Features

    • Introduced a new method enable_compression in the PairTabAtomicModel class, indicating that the model does not support compression settings.
  • Documentation

    • Added docstring for the enable_compression method to clarify its purpose.

Copy link
Contributor

coderabbitai bot commented Nov 27, 2024

📝 Walkthrough

Walkthrough

The changes introduce a new method called enable_compression to the PairTabAtomicModel class in two files: deepmd/dpmodel/atomic_model/pairtab_atomic_model.py and deepmd/pt/model/atomic_model/pairtab_atomic_model.py. This method accepts several parameters related to compression settings but does not implement any functionality, as it contains only a docstring and a pass statement. No other modifications to existing methods or attributes are included in this pull request.

Changes

File Path Change Summary
deepmd/dpmodel/atomic_model/pairtab_atomic_model.py Added method enable_compression with parameters for compression settings; contains a docstring and pass.
deepmd/pt/model/atomic_model/pairtab_atomic_model.py Added method enable_compression with parameters for compression settings; contains a docstring and pass.

Possibly related PRs

  • Enable Hybrid Descriptor to be compressed #4297: The enable_compression method added to the DescrptHybrid class shares the same method signature and purpose as the one introduced in the main PR, indicating a related enhancement in compression functionality.
  • Add compression API to BaseModel and AtomicModel #4298: The enable_compression method added to the DPAtomicModel class has the same parameters and intent as the main PR, suggesting a direct connection in the implementation of compression features across different models.
  • feat(pt): DPA-2 repinit compress #4329: The enable_compression method in the DescrptDPA2 class also follows the same pattern and parameter structure as the main PR, indicating a cohesive approach to adding compression capabilities across various descriptors.

Suggested reviewers

  • njzjz

📜 Recent review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between a9d6292 and e0d1085.

📒 Files selected for processing (2)
  • deepmd/dpmodel/atomic_model/pairtab_atomic_model.py (1 hunks)
  • deepmd/pt/model/atomic_model/pairtab_atomic_model.py (1 hunks)
✅ Files skipped from review due to trivial changes (2)
  • deepmd/dpmodel/atomic_model/pairtab_atomic_model.py
  • deepmd/pt/model/atomic_model/pairtab_atomic_model.py

Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?

❤️ Share
🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Generate unit testing code for this file.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query. Examples:
    • @coderabbitai generate unit testing code for this file.
    • @coderabbitai modularize this function.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read src/utils.ts and generate unit testing code.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.
    • @coderabbitai help me debug CodeRabbit configuration file.

Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments.

CodeRabbit Commands (Invoked using PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger an incremental review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai full review to do a full review from scratch and review all the files again.
  • @coderabbitai summary to regenerate the summary of the PR.
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai configuration to show the current CodeRabbit configuration for the repository.
  • @coderabbitai help to get help.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Documentation and Community

  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 2

🧹 Outside diff range and nitpick comments (5)
deepmd/dpmodel/atomic_model/linear_atomic_model.py (4)

Line range hint 442-445: Improve assertion message clarity

The assertion message could be more descriptive to help developers understand the configuration requirements better.

-   assert (
-       self.sw_rmax > self.sw_rmin
-   ), "The upper boundary `sw_rmax` must be greater than the lower boundary `sw_rmin`."
+   assert (
+       self.sw_rmax > self.sw_rmin
+   ), f"Invalid boundary configuration: sw_rmax ({self.sw_rmax}) must be greater than sw_rmin ({self.sw_rmin})"

Line range hint 474-481: Handle division by zero more explicitly

The current implementation uses numpy's error state management to handle division by zero, but this could be more explicit and safer.

Consider this approach:

-   with np.errstate(divide="ignore", invalid="ignore"):
-       sigma = numerator / denominator
+   # Avoid division by zero
+   mask = denominator != 0
+   sigma = np.zeros_like(numerator)
+   sigma[mask] = numerator[mask] / denominator[mask]

Line range hint 482-492: Consider extracting smoothing function

The smoothing calculation is complex and could benefit from being extracted into a separate method for better readability and maintainability.

Consider extracting the smoothing logic:

+   def _compute_smooth_coefficient(self, u: np.ndarray) -> np.ndarray:
+       """Compute smooth interpolation coefficient.
+       
+       Parameters
+       ----------
+       u : np.ndarray
+           Normalized distance values
+       
+       Returns
+       -------
+       np.ndarray
+           Smooth interpolation coefficient
+       """
+       return -6 * u**5 + 15 * u**4 - 10 * u**3 + 1

    def _compute_weight(self, ...):
        # ... existing code ...
        with np.errstate(invalid="ignore"):
-           smooth = -6 * u**5 + 15 * u**4 - 10 * u**3 + 1
+           smooth = self._compute_smooth_coefficient(u)

Line range hint 493-496: Document the weight calculation logic

The weight calculation logic is complex but lacks documentation explaining the mathematical reasoning behind the coefficients.

Add a docstring section explaining the mathematical formula:

    def _compute_weight(self, ...):
        """ZBL weight.
+
+       The weight is calculated using a smooth step function:
+       1. For distances < sw_rmin: weight = 1
+       2. For distances > sw_rmax: weight = 0
+       3. For distances in between: weight = smooth polynomial interpolation
+          using a 5th-degree polynomial: -6x^5 + 15x^4 - 10x^3 + 1
+
+       This ensures continuous first and second derivatives at the boundaries.

        Returns
        -------
        list[np.ndarray]
            the atomic ZBL weight for interpolation. (nframes, nloc, 1)
        """
deepmd/pt/model/atomic_model/linear_atomic_model.py (1)

Line range hint 588-652: Improve documentation for better maintainability

The weight computation logic is complex but lacks detailed documentation. Consider:

  1. Adding mathematical formulas in docstring to explain the interpolation
  2. Documenting the meaning of variables (u, sigma, coef)
  3. Explaining the choice of the polynomial coefficients (-6, 15, -10)

Add this documentation:

 def _compute_weight(
     self,
     extended_coord: torch.Tensor,
     extended_atype: torch.Tensor,
     nlists_: list[torch.Tensor],
 ) -> list[torch.Tensor]:
-    """ZBL weight.
+    """Compute interpolation weights between DP and ZBL models.
+    
+    The interpolation uses a smooth polynomial function:
+    - For r < sw_rmin: weight_zbl = 1
+    - For sw_rmin <= r < sw_rmax: weight_zbl = -6u⁵ + 15u⁴ - 10u³ + 1
+      where u = (r - sw_rmin)/(sw_rmax - sw_rmin)
+    - For r >= sw_rmax: weight_zbl = 0
+    
+    The distance r is computed as a softmin over all neighbor distances:
+    r = Σ(d_i * exp(-d_i/α)) / Σ(exp(-d_i/α))
+    where d_i are pairwise distances and α is smin_alpha.
 
     Returns
     -------
     list[torch.Tensor]
         the atomic ZBL weight for interpolation. (nframes, nloc, 1)
     """
📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between f343a3b and a9d6292.

📒 Files selected for processing (2)
  • deepmd/dpmodel/atomic_model/linear_atomic_model.py (1 hunks)
  • deepmd/pt/model/atomic_model/linear_atomic_model.py (1 hunks)
🔇 Additional comments (1)
deepmd/pt/model/atomic_model/linear_atomic_model.py (1)

577-587: Verify compression method usage across codebase

Let's ensure that no code relies on the compression functionality for PairTab models.

✅ Verification successful

Compression is correctly handled for PairTab models

The verification shows that the implementation is consistent across the codebase:

  • The enable_compression method is correctly implemented as a no-op with a docstring explicitly stating "Pairtab model does not support compression" in both PyTorch and TensorFlow implementations
  • No code attempts to use compression functionality with PairTab models
  • The compression-related configuration and documentation properly handle this limitation
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Search for calls to enable_compression on PairTab instances
rg -l "PairTabAtomicModel" | xargs rg "enable_compression"

# Search for compression-related configuration
rg -l "pairtab" | xargs rg -i "compress"

Length of output: 3014

deepmd/dpmodel/atomic_model/linear_atomic_model.py Outdated Show resolved Hide resolved
deepmd/pt/model/atomic_model/linear_atomic_model.py Outdated Show resolved Hide resolved
@anyangml anyangml linked an issue Nov 27, 2024 that may be closed by this pull request
@anyangml anyangml requested a review from njzjz November 27, 2024 04:42
Copy link

codecov bot commented Nov 27, 2024

Codecov Report

Attention: Patch coverage is 50.00000% with 2 lines in your changes missing coverage. Please review.

Project coverage is 84.64%. Comparing base (f343a3b) to head (e0d1085).
Report is 1 commits behind head on devel.

Files with missing lines Patch % Lines
...eepmd/dpmodel/atomic_model/pairtab_atomic_model.py 50.00% 1 Missing ⚠️
...epmd/pt/model/atomic_model/pairtab_atomic_model.py 50.00% 1 Missing ⚠️
Additional details and impacted files
@@            Coverage Diff             @@
##            devel    #4432      +/-   ##
==========================================
- Coverage   84.64%   84.64%   -0.01%     
==========================================
  Files         614      614              
  Lines       57138    57141       +3     
  Branches     3487     3486       -1     
==========================================
  Hits        48367    48367              
- Misses       7646     7647       +1     
- Partials     1125     1127       +2     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@anyangml anyangml added this pull request to the merge queue Nov 27, 2024
Merged via the queue into deepmodeling:devel with commit 3cdf407 Nov 27, 2024
60 checks passed
@anyangml anyangml deleted the feat/add-pairtab-compression branch November 27, 2024 09:05
@coderabbitai coderabbitai bot mentioned this pull request Dec 1, 2024
9 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

The torch backend compress error
2 participants