ICONN AI Logo

ICONN 1

Introducing ICONN 1 Mini Beta, a cutting-edge open-source AI model with just 7 billion parameters — designed for natural, human-like language understanding and generation. Despite its compact size, it delivers powerful performance through efficient architecture and careful tuning. ICONN 1 Mini Beta represents the next step in accessible, conversational AI.

Developed entirely from scratch, ICONN-1-Mini-Beta is based on a new ICONN framework and comprises 7 billion parameters.

ICONN-1 is released in three distinct forms to serve different application needs:

  • ICONN-1-Mini-Beta(This model) is a small 7B model trained for a lightweight alternative to ICONN 1.
  • ICONN-1 is optimized for natural, emotionally resonant, and conversational interactions.
  • ICONN-e1 is a specialized variant of the model fine-tuned for advanced reasoning, critical analysis, and complex problem-solving.

Together, these models represent a major leap forward in the evolution of AI systems—demonstrating not only deep reasoning but also a commitment to openness, accessibility, and human-aligned intelligence.

Usage

To run ICONN 1 Mini Beta, you need:

  • Any hardware - CPU or GPU; Just make sure you have about 15GB storage space!

Run the code below to run ICONN 1 Mini Beta:

import os

import torch

from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer

from threading import Thread

  

model_id = "ICONNAI/ICONN-1-Mini-Beta"

  

try:

model = AutoModelForCausalLM.from_pretrained(

model_id, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True

)

tokenizer = AutoTokenizer.from_pretrained(model_id)

except Exception as e:

exit(f"Exiting due to model loading error: {e}")

  

def generate_response(

message:  str,

max_new_tokens:  int = 2048,

temperature:  float = 0.7,

top_p:  float = 0.9,

top_k:  int = 50,

repetition_penalty:  float = 1.2,

) -> str:

conversation = [{"role":  "user",  "content": message}]

  

try:

input_ids = tokenizer.apply_chat_template(

conversation, return_tensors="pt", enable_thinking=True

)

except Exception as e:

return  f"Error applying chat template: {e}"

  

input_ids = input_ids.to(model.device)

  

streamer = TextIteratorStreamer(tokenizer, timeout=20.0, skip_prompt=True, skip_special_tokens=True)

  

adjusted_top_k = int(max(1, top_k))

  

generate_kwargs = dict(

{"input_ids": input_ids},

streamer=streamer,

max_new_tokens=max_new_tokens,

do_sample=True,

top_p=top_p,

top_k=adjusted_top_k,

temperature=temperature,

num_beams=1,

repetition_penalty=repetition_penalty,

)

  

try:

t = Thread(target=model.generate, kwargs=generate_kwargs)

t.start()

except Exception as e:

return  f"Error starting generation thread: {e}"

  

outputs = []

for text in streamer:

outputs.append(text)

return  "".join(outputs)

  

if  __name__ == "__main__":

question = "Can you explain briefly to me what is the Python programming language?"

print(f"User Question: {question}")

  

response = generate_response(question)

print(f"Bot Response: {response}")
Downloads last month
27
Safetensors
Model size
7.83B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Spaces using ICONNAI/ICONN-1-Mini-Beta 2