Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
Omar Sanseviero a24194737d | 1 month ago | |
---|---|---|
colabs | 1 month ago | |
examples | 2 months ago | |
gemma | 1 month ago | |
CONTRIBUTING.md | 2 months ago | |
LICENSE | 2 months ago | |
README.md | 1 month ago | |
pyproject.toml | 1 month ago |
Gemma is a family of open-weights Large Language
Model (LLM) by Google DeepMind, based on Gemini
research and technology.
This repository contains an inference implementation and examples, based on the
Flax and JAX.
To install Gemma you need to use Python 3.10 or higher.
Install JAX for CPU, GPU or TPU. Follow instructions at
the JAX website.
Run
python -m venv gemma-demo
. gemma-demo/bin/activate
pip install git+https://github.com/google-deepmind/gemma.git
The model checkpoints are available through Kaggle at
http://kaggle.com/models/google/gemma. Select one of the Flax model
variations, click the ⤓ button to download the model archive, then extract the
contents to a local directory.
Alternatively, visit the gemma
models on the Hugging Face Hub. To download the model, you can run the following code if you have huggingface_hub
installed:
from huggingface_hub import snapshot_download
local_dir = snapshot_download(repo_id="google/gemma-2b-flax")
snapshot_download(repo_id="google/gemma-2b-flax", local_dir=local_dir)
In both cases, the archive contains both the model weights and
the tokenizer, for example the 2b Flax variation contains:
2b/ # Directory containing model weights
tokenizer.model # Tokenizer
To run the unit tests, install the optional [test]
dependencies (e.g. using
pip install -e .[test]
from the root of the source tree), then:
pytest .
Note that the tests in sampler_test.py
are skipped by default since no
tokenizer is distributed with the Gemma sources. To run these tests, download a
tokenizer following the instructions above, and update the _VOCAB
constant in
sampler_test.py
with the path to tokenizer.model
.
To run the example sampling script, pass the paths to the weights directory and
tokenizer:
python examples/sampling.py \
--path_checkpoint=/path/to/archive/contents/2b/ \
--path_tokenizer=/path/to/archive/contents/tokenizer.model
There are also several Colab notebook tutorials:
colabs/sampling_tutorial.ipynb
contains a Colab notebook with a sampling example.
colabs/fine_tuning_tutorial.ipynb
contains a Colab with a basic tutorial on how to fine
tune Gemma for a task, such as English to French translation.
colabs/gsm8k_eval.ipynb
is a Colab with a reference GSM8K eval
implementation.
To run these notebooks you will need to download a local copy of the weights and
tokenizer (see above), and update the ckpt_path
and vocab_path
variables
with the corresponding paths.
Gemma can run on a CPU, GPU and TPU. For GPU, we recommend a 8GB+ RAM on GPU for
the 2B checkpoint and 24GB+ RAM on GPU for the 7B checkpoint.
We are open to bug reports, pull requests (PR), and other contributions. Please
see CONTRIBUTING.md for details on PRs.
Copyright 2024 DeepMind Technologies Limited
This code is licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License. You may obtain
a copy of the License at http://www.apache.org/licenses/LICENSE-2.0.
Unless required by applicable law or agreed to in writing, software distributed
under the License is distributed on an AS IS BASIS, WITHOUT WARRANTIES OR
CONDITIONS OF ANY KIND, either express or implied. See the License for the
specific language governing permissions and limitations under the License.
This is not an official Google product.
No Description
Jupyter Notebook Python
Dear OpenI User
Thank you for your continuous support to the Openl Qizhi Community AI Collaboration Platform. In order to protect your usage rights and ensure network security, we updated the Openl Qizhi Community AI Collaboration Platform Usage Agreement in January 2024. The updated agreement specifies that users are prohibited from using intranet penetration tools. After you click "Agree and continue", you can continue to use our services. Thank you for your cooperation and understanding.
For more agreement content, please refer to the《Openl Qizhi Community AI Collaboration Platform Usage Agreement》