runtime error

Space failed. Exit code: 1. Reason: 00, 102MB/s] Downloading pytorch_model.bin: 99%|█████████▉| 3.85G/3.89G [00:38<00:00, 102MB/s] Downloading pytorch_model.bin: 99%|█████████▉| 3.86G/3.89G [00:38<00:00, 98.6MB/s] Downloading pytorch_model.bin: 100%|█████████▉| 3.88G/3.89G [00:38<00:00, 101MB/s] Downloading pytorch_model.bin: 100%|█████████▉| 3.89G/3.89G [00:38<00:00, 101MB/s] Downloading pytorch_model.bin: 100%|██████████| 3.89G/3.89G [00:38<00:00, 100MB/s] No compiled kernel found. Compiling kernels : /home/user/.cache/huggingface/modules/transformers_modules/THUDM/chatglm-6b-int4/6c5205c47d0d2f7ea2e44715d279e537cae0911f/quantization_kernels_parallel.c Compiling gcc -O3 -fPIC -pthread -fopenmp -std=c99 /home/user/.cache/huggingface/modules/transformers_modules/THUDM/chatglm-6b-int4/6c5205c47d0d2f7ea2e44715d279e537cae0911f/quantization_kernels_parallel.c -shared -o /home/user/.cache/huggingface/modules/transformers_modules/THUDM/chatglm-6b-int4/6c5205c47d0d2f7ea2e44715d279e537cae0911f/quantization_kernels_parallel.so Load kernel : /home/user/.cache/huggingface/modules/transformers_modules/THUDM/chatglm-6b-int4/6c5205c47d0d2f7ea2e44715d279e537cae0911f/quantization_kernels_parallel.so Setting CPU quantization kernel threads to 8 Using quantization cache Applying quantization to glm layers Traceback (most recent call last): File "app.py", line 99, in <module> emptyBtn.click(reset_state, outputs=[chatbot, history], show_progress=True) File "/home/user/.local/lib/python3.8/site-packages/gradio/blocks.py", line 1200, in __exit__ self.config = self.get_config_file() File "/home/user/.local/lib/python3.8/site-packages/gradio/blocks.py", line 1176, in get_config_file "input": list(block.input_api_info()), # type: ignore File "/home/user/.local/lib/python3.8/site-packages/gradio_client/serializing.py", line 41, in input_api_info return (api_info["serialized_input"][0], api_info["serialized_input"][1]) KeyError: 'serialized_input'

Container logs:

Fetching error logs...