← Back to Product Feed

GitHub Open Source fikrikarim/parlor

On-device, real-time multimodal AI. Have natural voice and vision conversations with an AI that runs entirely on your machine. Powered by Gemma 4 E2B and Kokoro.

1,576
Traction Score
176
Forks
Apr 5, 2026
Launch Date
View Origin Link

Product Positioning & Context

On-device, real-time multimodal AI. Have natural voice and vision conversations with an AI that runs entirely on your machine. Powered by Gemma 4 E2B and Kokoro.
apple-silicon gemma kokoro litert-lm local-llm mlx multimodal on-device-ai

Related Ecosystem & Alternatives

Discover adjacent products, open-source repositories, and developer tools sharing similar technical architecture.

Deep-Dive FAQs

What is fikrikarim/parlor?
fikrikarim/parlor is a digital product or tool described as: On-device, real-time multimodal AI. Have natural voice and vision conversations with an AI that runs entirely on your machine. Powered by Gemma 4 E...
Where did fikrikarim/parlor originate?
Data for fikrikarim/parlor was aggregated directly from the GitHub Open Source community ecosystem, representing raw developer and early-adopter sentiment.
When was fikrikarim/parlor publicly launched?
The initial public indexing or launch date for fikrikarim/parlor within our tracked developer communities was recorded on April 5, 2026.
How popular is fikrikarim/parlor?
fikrikarim/parlor has achieved measurable traction, logging over 1,576 traction score and facilitating 176 recorded discussions or engagements.
Which technical categories define fikrikarim/parlor?
Based on metadata extraction, fikrikarim/parlor is categorized under topics such as: apple-silicon, gemma, kokoro, litert-lm.
Are there active development issues for fikrikarim/parlor?
Yes, we are currently tracking open architectural debates and bug reports for this project on GitHub. There are currently 1 active high-priority issues logged recently.
How does the creator describe fikrikarim/parlor?
The original author or development team describes the product as follows: "On-device, real-time multimodal AI. Have natural voice and vision conversations with an AI that runs entirely on your machine. Powered by Gemma 4 E2B and Kokoro."

Active Developer Issues (GitHub)

open Running on WSL2 Ubuntu 22.04 fails
Logged: Apr 6, 2026

Community Voice & Feedback

yhdanid • Apr 14, 2026
I have looked around briefly concerning this, and it seems the matter is not only related to litert (which seems to rely on Vulkan for GPU hardware acceleration), but I've also found out that even llama.cpp's Vulkan implementation won't use the dedicated NVIDIA GPU in WSL. Even when I specify --device MY_VULKAN_DEVICE (for NVIDIA) in the command. It still falls back to CPU acceleration. So this is not a litert issue, or llama.cpp issue, or a Parlor issue, but rather a Vulkan issue when running in WSL.

So I suggest closing this as there is nothing that can be done until a solution upstream is found for Vulkan running in WSL in GPU acceleration mode.
fikrikarim • Apr 11, 2026
Thanks for the additional information.

Could you changing the `litert_lm.Backend` to be CPU on all backend on the `server.py`?
```python
engine = litert_lm.Engine(
MODEL_PATH,
backend=litert_lm.Backend.CPU,
vision_backend=litert_lm.Backend.CPU,
audio_backend=litert_lm.Backend.CPU,
)
```

If that works, then there's a problem with the GPU inference on your setup.

If you could isolate and reproduce the issue just on python `litert_lm` without additional stuff from this repo, perhaps it'd be better to report the issue or ask for support from the main [LiteRT-LM repo](https://github.com/google-ai-edge/LiteRT-LM)?
yhdanid • Apr 7, 2026
### SUMMARY
I've tried with --backend=cpu, it works fine, and also seems to work when --backend=gpu is specified. Benchmarking fails on GPU backend but works fine in CPU. So I'm confused.

Note that I did this with model locally downloaded in /model directory (to save on bandwidth as I was testing), and while downloading from huggingface. All this was done in WSL2, Ubuntu 22.04.5 LTS.

**GPU IS ACCESSIBLE FROM TORCH**
As an experiment, I installed pytorch in another virtual environment and tested if it can access the GPU, and I got these results:

```
(.venv) nemisis@nemisis-BLACK:/mnt/e/sandbox/parlor/test-cuda$ python torch_cuda_test.py
Is CUDA available? True
CUDA device count: 1
CUDA current device: 0
Torch CUDA device:
CUDA device name: NVIDIA GeForce RTX 4080 GPU
```

**DOCKER**
I also tried seeing if parlor works in docker as it is more widely available than WSL (image built FROM nvidia/cuda:12.8.0-devel-ubuntu24.04) with more or les...
fikrikarim • Apr 7, 2026
@yhdanid could you try running the `litert-lm` CLI with the gpu backend? The default backend is CPU.
```
litert-lm run \
--backend=gpu \
--from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \
gemma-4-E2B-it.litertlm \
--prompt="What is the capital of France?"
```

If that works, try running the `litert-lm benchmark` to double check whether it's actually running on the GPU:
```
litert-lm benchmark \
--backend=gpu \
--from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \
gemma-4-E2B-it.litertlm
```

and compare with
```
litert-lm benchmark \
--backend=cpu \
--from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \
gemma-4-E2B-it.litertlm
```
fikrikarim • Apr 6, 2026
Thanks for the detailed information. Unfortunately, I don't have a Windows machine myself so it's hard for me to debug it. I'll give it a try since there seems to be some people that show interest on running this on Windows. Although, I can't promise anything right now.

Discovery Source

GitHub Open Source GitHub Open Source

Aggregated via automated community intelligence tracking.

Tech Stack Dependencies

No direct open-source NPM package mentions detected in the product documentation.

Media Tractions & Mentions

No mainstream media stories specifically mentioning this product name have been intercepted yet.

Deep Research & Science

No direct peer-reviewed scientific literature matched with this product's architecture.