Spaces:
				
			
			
	
			
			
		Sleeping
		
	
	
	
			
			
	
	
	
	
		
		
		Sleeping
		
	Commit 
							
							·
						
						006bcec
	
1
								Parent(s):
							
							14ad06d
								
fix: docker arg step
Browse files- cloudbuild.yaml +5 -1
- core-model-prediction/Dockerfile +2 -5
    	
        cloudbuild.yaml
    CHANGED
    
    | @@ -6,7 +6,7 @@ steps: | |
| 6 | 
             
                args: 
         | 
| 7 | 
             
                  - -c
         | 
| 8 | 
             
                  - | 
         | 
| 9 | 
            -
                      docker build -t us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest --build-arg | 
| 10 |  | 
| 11 | 
             
              - name: "gcr.io/cloud-builders/docker"
         | 
| 12 | 
             
                args:
         | 
| @@ -15,6 +15,10 @@ steps: | |
| 15 | 
             
                    "us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest",
         | 
| 16 | 
             
                  ]
         | 
| 17 |  | 
|  | |
|  | |
|  | |
|  | |
| 18 | 
             
            images:
         | 
| 19 | 
             
              - "us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest"
         | 
| 20 |  | 
|  | |
| 6 | 
             
                args: 
         | 
| 7 | 
             
                  - -c
         | 
| 8 | 
             
                  - | 
         | 
| 9 | 
            +
                      docker build -t us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest --build-arg HF_TOKEN=$$_HF_TOKEN .
         | 
| 10 |  | 
| 11 | 
             
              - name: "gcr.io/cloud-builders/docker"
         | 
| 12 | 
             
                args:
         | 
|  | |
| 15 | 
             
                    "us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest",
         | 
| 16 | 
             
                  ]
         | 
| 17 |  | 
| 18 | 
            +
            options:
         | 
| 19 | 
            +
              pool:
         | 
| 20 | 
            +
                name: 'projects/${PROJECT_ID}/locations/us-central1/workerPools/ai-detector-builder'
         | 
| 21 | 
            +
             | 
| 22 | 
             
            images:
         | 
| 23 | 
             
              - "us-central1-docker.pkg.dev/${PROJECT_ID}/interview-ai-detector/model-prediction:latest"
         | 
| 24 |  | 
    	
        core-model-prediction/Dockerfile
    CHANGED
    
    | @@ -1,6 +1,3 @@ | |
| 1 | 
            -
            # HF Token args
         | 
| 2 | 
            -
            ARG HF_TOKEN
         | 
| 3 | 
            -
             | 
| 4 | 
             
            # Use an official Python runtime as a base image
         | 
| 5 | 
             
            FROM pytorch/pytorch:2.1.2-cuda12.1-cudnn8-runtime
         | 
| 6 |  | 
| @@ -20,8 +17,8 @@ RUN python -m nltk.downloader punkt wordnet averaged_perceptron_tagger | |
| 20 | 
             
            # Unzip wordnet
         | 
| 21 | 
             
            RUN unzip /root/nltk_data/corpora/wordnet.zip -d /root/nltk_data/corpora/
         | 
| 22 |  | 
| 23 | 
            -
            #  | 
| 24 | 
            -
             | 
| 25 |  | 
| 26 | 
             
            # Download HuggingFace model
         | 
| 27 | 
             
            RUN python -c "from transformers import AutoTokenizer, AutoModelForCausalLM; \
         | 
|  | |
|  | |
|  | |
|  | |
| 1 | 
             
            # Use an official Python runtime as a base image
         | 
| 2 | 
             
            FROM pytorch/pytorch:2.1.2-cuda12.1-cudnn8-runtime
         | 
| 3 |  | 
|  | |
| 17 | 
             
            # Unzip wordnet
         | 
| 18 | 
             
            RUN unzip /root/nltk_data/corpora/wordnet.zip -d /root/nltk_data/corpora/
         | 
| 19 |  | 
| 20 | 
            +
            # HF Token args
         | 
| 21 | 
            +
            ARG HF_TOKEN
         | 
| 22 |  | 
| 23 | 
             
            # Download HuggingFace model
         | 
| 24 | 
             
            RUN python -c "from transformers import AutoTokenizer, AutoModelForCausalLM; \
         |