finagg
trl
finagg | trl | |
---|---|---|
17 | 13 | |
387 | 8,291 | |
- | 6.2% | |
8.1 | 9.7 | |
8 days ago | 5 days ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
finagg
-
This Week In Python
finagg – A Python package for aggregating and normalizing historical data from popular and free financial APIs
- FLaNK Stack 29 Jan 2024
- Show HN: Finagg – free and nearly unlimited financial data
-
[D] Website to get historical price for agriculture commodities?
This is certainly a weird place to ask this question. That being said, you should explore the FRED API. Here's my project that implements most of it in Python: https://github.com/theOGognf/finagg The walkthrough shows you how to find what you're looking for
-
Fundamental Data Sources
I created a Python package exactly for this. https://github.com/theOGognf/finagg. It aggregates historical fundamental data for whatever tickers you specify or from a subset of tickers. Let me know what you think
-
Is accurate quarterly earnings data availible?
This package https://github.com/theOGognf/finagg already implements the complete SEC EDGAR REST API (disclaimer: I'm the author), and the archive-based API is in the works. I suggest you give it a go using the latest version off GitHub
-
Sunday Daily Thread: What's everyone working on this week?
I've got some time set aside to implement a (file based) SEC EDGAR API described in this issue https://github.com/theOGognf/finagg/issues/43
- finagg: NEW Data - star count:107.0
trl
- FLaNK Stack 29 Jan 2024
-
OOM Error while using TRL for RLHF Fine-tuning
I am using TRL for RLHF fine-tuning the Llama-2-7B model and getting an OOM error (even with batch_size=1). If anyone used TRL for RLHF can please tell me what I am doing wrong? Code details can be found in the GitHub issue.
-
[D] Tokenizers Truncation during Fine-tuning with Large Texts
SFTtrainer from huggingface
-
New Open-source LLMs! 🤯 The Falcon has landed! 7B and 40B
For lora - PEFT seems to work. I don't have patience to wait 5 hours, but modifying this example seems to work. You don't even need to modify that much, as their model just as neo-x uses query_key_value name for self-attention.
-
[D] Using RLHF beyond preference tuning
They have examples of making GPT output more positive (code) by using a sentiment model as reward. There are other examples about reducing toxicity, summarization here: https://github.com/lvwerra/trl/tree/main/examples . Should be fairly simple to modify the sentiment example and try the calculator reward you mentioned above.
-
[R] 🤖🌟 Unlock the Power of Personal AI: Introducing ChatLLaMA, Your Custom Personal Assistant! 🚀💬
You can use this -> https://github.com/lvwerra/trl/blob/main/examples/sentiment/scripts/gpt-neox-20b_peft/merge_peft_adapter.py
-
[R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003
Just the hh directly. From the results it seems like it might possibly be enough but I might also try instruction tuning then running the whole process from that base. I will also be running the reinforcement learning by using a Lora using this as an example https://github.com/lvwerra/trl/tree/main/examples/sentiment/scripts/gpt-neox-20b_peft
-
[R] A simple explanation of Reinforcement Learning from Human Feedback (RLHF)
This package is pretty simple to use! https://github.com/lvwerra/trl
- Transformer Reinforcement Learning
- trl: Train transformer language models with reinforcement learning
What are some alternatives?
pyautoenv - Automatically activate and deactivate Python environments as you move around the file system.
lm-human-preferences - Code for the paper Fine-Tuning Language Models from Human Preferences
sql-to-kml - Format SQL query results into a KML file.
alpaca-lora - Instruct-tune LLaMA on consumer hardware
plombery - Python task scheduler with a user-friendly web UI
trlx - A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
usepython - Run Python scripts in a Pyodide service worker
LLaMA-8bit-LoRA - Repository for Chat LLaMA - training a LoRA for the LLaMA (1 or 2) models on HuggingFace with 8-bit or 4-bit quantization. Research only.
bytewax - Python Stream Processing
sparsegpt-for-LLaMA - Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.
java-snapshot-testing - Facebook style snapshot testing for JAVA Tests
llama-recipes - Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.