Skip to content

Conversation

@puririshi98
Copy link
Contributor

@puririshi98 puririshi98 commented Jun 26, 2024

@puririshi98 puririshi98 self-assigned this Jun 26, 2024
@github-actions github-actions bot added the nn label Jun 26, 2024
@codecov
Copy link

codecov bot commented Jun 26, 2024

Codecov Report

Attention: Patch coverage is 15.27778% with 122 lines in your changes missing coverage. Please review.

Project coverage is 87.50%. Comparing base (dafbd30) to head (5c7e7b2).
Report is 32 commits behind head on master.

Current head 5c7e7b2 differs from pull request most recent head 3d9e354

Please upload reports for the commit 3d9e354 to get more accurate results.

Files Patch % Lines
torch_geometric/nn/nlp/llm.py 15.27% 122 Missing ⚠️
Additional details and impacted files
@@            Coverage Diff             @@
##           master    #9462      +/-   ##
==========================================
- Coverage   88.25%   87.50%   -0.75%     
==========================================
  Files         473      476       +3     
  Lines       30844    31120     +276     
==========================================
+ Hits        27221    27233      +12     
- Misses       3623     3887     +264     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@rusty1s rusty1s changed the title breaking PR 9167 into smaller PRs. Starting w/ adding LLM Integrate LLM class Jul 2, 2024
Copy link
Member

@rusty1s rusty1s left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Very cool :) Made some formatting changes, and inference method needed the bos_token_id to not crash. Please confirm :)

@rusty1s rusty1s merged commit f0ef2de into master Jul 2, 2024
@rusty1s rusty1s deleted the llm-integ branch July 2, 2024 12:16
rusty1s added a commit that referenced this pull request Sep 10, 2024
1. #9462
2. **->** #9480
3. #9481
4. #9167

---

breaking #9167 down
further, focusing on G-retriever model this time

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Akihiro Nitta <[email protected]>
Co-authored-by: rusty1s <[email protected]>
rusty1s added a commit that referenced this pull request Sep 13, 2024
1. #9462
2. #9480
3. **->** #9481
4. #9167

---

Breaking down PR
#9167 further

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Akihiro Nitta <[email protected]>
Co-authored-by: rusty1s <[email protected]>
rusty1s added a commit that referenced this pull request Sep 13, 2024
1. #9462
2. #9480
3. #9481
4. **->** #9167

---

repro:
Latest NVIDIA PyG container
+
`git config --global credential.helper store; huggingface-cli login; cd
/opt/pyg; pip uninstall -y torch-geometric; rm -rf pytorch_geometric;
git clone -b gnn-llm-model-integration
https://github.com/pyg-team/pytorch_geometric.git; cd
/opt/pyg/pytorch_geometric; pip install .; pip install peft datasets
transformers pcst_fast sentencepiece; python3
examples/llm_plus_gnn/g_retriever.py`

old PR: #9154

note: pure cpu is 220x slower than pure GPU using a single Grace Hopper
(for llama-7b)

info:
tried gemma, performs worse in all train/val/test metrics. most likely
needs some tuning, will leave this as future work as part of the
community sprint to try many LLM and GNN combos and tune them. Therefore
keeping the default llama2

the new gemma-v2 is also much worse than llama2

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Akihiro Nitta <[email protected]>
Co-authored-by: rusty1s <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants