Upload google_gemma-3-27b-it_2.py with huggingface_hub
Browse files
google_gemma-3-27b-it_2.py
CHANGED
|
@@ -21,10 +21,10 @@
|
|
| 21 |
|
| 22 |
try:
|
| 23 |
# Load model directly
|
| 24 |
-
from transformers import AutoProcessor,
|
| 25 |
|
| 26 |
processor = AutoProcessor.from_pretrained("google/gemma-3-27b-it")
|
| 27 |
-
model =
|
| 28 |
messages = [
|
| 29 |
{
|
| 30 |
"role": "user",
|
|
@@ -59,10 +59,10 @@ except Exception as e:
|
|
| 59 |
import traceback
|
| 60 |
f.write('''```CODE:
|
| 61 |
# Load model directly
|
| 62 |
-
from transformers import AutoProcessor,
|
| 63 |
|
| 64 |
processor = AutoProcessor.from_pretrained("google/gemma-3-27b-it")
|
| 65 |
-
model =
|
| 66 |
messages = [
|
| 67 |
{
|
| 68 |
"role": "user",
|
|
|
|
| 21 |
|
| 22 |
try:
|
| 23 |
# Load model directly
|
| 24 |
+
from transformers import AutoProcessor, AutoModelForMultimodalLM
|
| 25 |
|
| 26 |
processor = AutoProcessor.from_pretrained("google/gemma-3-27b-it")
|
| 27 |
+
model = AutoModelForMultimodalLM.from_pretrained("google/gemma-3-27b-it")
|
| 28 |
messages = [
|
| 29 |
{
|
| 30 |
"role": "user",
|
|
|
|
| 59 |
import traceback
|
| 60 |
f.write('''```CODE:
|
| 61 |
# Load model directly
|
| 62 |
+
from transformers import AutoProcessor, AutoModelForMultimodalLM
|
| 63 |
|
| 64 |
processor = AutoProcessor.from_pretrained("google/gemma-3-27b-it")
|
| 65 |
+
model = AutoModelForMultimodalLM.from_pretrained("google/gemma-3-27b-it")
|
| 66 |
messages = [
|
| 67 |
{
|
| 68 |
"role": "user",
|