Skip to content

Silk: AI-Powered Task Automation

Silk is a powerful terminal tool for quick task automation using Large Language Models (LLMs). It ships as a single executable and works best with Silky.dev, or any OpenAI-compatible provider.

What is Silk?

Silk enables you to generate code, write articles, books, websites, and more through:

  • One-off commands: Quick, direct task execution
  • Interactive chat mode: Gradual task development

What sets Silk apart is its performance and versatility.

Use Cases

Silk is the ultimate task automator. Examples include:

  • Program Creation: Generate code without manual writing

    silk run 'create a snake game'
  • Content Generation:

    • Write summaries
    • Create websites
    • Generate documentation
  • Code Conversion: Seamlessly transform between frameworks

    silk run 'convert this react code to vue'

Performance Optimization

Our core focus is efficiency in AI computation. As AI compute is energy hungry our goal is to utilize this as efficient as possible. This greatly reduces the cost and as bonus leads to fast execution. We do this by:

  • Auto Model Selection: automatically selecting the optimal model for a given task
  • Multi-Step Calling: Break complex tasks into efficient smaller calls
  • Intelligent Caching: Reuse and optimize previous computations

Note: Advanced features are available exclusively with silky.dev

Getting Started

Online Version

Local Usage

Works offline with reduced features:

sh
# Specify provider and model
silk --provider ollama --model=llama3.1
silk --provider lmstudio --model=llama3.1

# Shorthand for local Ollama/Llama3.1
silk --local

FAQ

Which providers are available?

  • Silky.dev: Default provider with full feature set
  • OpenAI-Compatible Providers: Limited features
    • Recommended: Smart reasoning models like Llama3.1 7B