Show HN: I built an ISP infrastructure emulator from scratch with a custom vBNG
Positioned as a learning reference and starting point for individuals struggling to understand complex ISP networking, particularly BNGs and subscriber management, due to closed-source vendor stacks and lack of mentorship. It is explicitly stated as not production-grade, but an educational tool.
View Origin LinkProduct Positioning & Context
Developers and network engineers will find this project invaluable for several reasons. It offers a transparent, hands-on platform to dissect and understand the intricate mechanics of ISP infrastructure, moving beyond abstract concepts. The use of Containerlab for network topology emulation, coupled with a Python-based vBNG, demonstrates how modern software development practices and containerization can be applied to complex networking problems. This fosters experimentation, allows for prototyping custom solutions, and bridges the gap between traditional network engineering and software development. The project represents a broader trend towards open-source alternatives in specialized domains, the virtualization of complex infrastructure for learning and development, and the application of modern software architectures (like event-driven systems with Redis Streams) to networking. While not production-ready, Aether serves as a critical educational tool, empowering a new generation of network professionals to innovate and understand beyond vendor lock-in.
GitHub: https://github.com/saphalpdyl/AetherAether is a multi-BNG (Broadband Network Gateway) ISP infrastructure lab built almost from scratch that emulates IPoE IPv4 subscriber management end-to-end. It supports IPoE/Ipv4 networks and runs a python-based vBNG with RADIUS AAA, per-subscriber traffic shaping, and traffic simulation emulated on Containerlab. It is also my first personal networking project, built roughly over a month.Motivations behind the projectI'm a CS sophomore. About three years ago, I was assigned, as an intern, to build a OSS/BSS platform for a regional ISP by myself without mentoring. Referencing demo.splynx.com , I developed most of the BSS side ( bookkeeping, accounting, inventory management ), but, in terms of networking, I managed to install and setup RADIUS and that was about it. I didn't have anyone to mentor me or ask questions to, so I had given up then.Three years later, I decided to try cracking it again. This project is meant to serve as a learning reference for anyone who's been in that same position i.e staring at closed-source vendor stacks without proper guidance. This is absolutely not production-grade, but I hope it gives someone a place to start.Architecture overviewThe core component, the BNG, runs on an event-driven architecture where state changes are passed around as messages to avoid handling mutexes and locks. The session manager is the sole owner of the session state. To keep it clean and predictable, the direBNG never accepts external inputctly. The one exception is the Go RADIUS CoA daemon, which passes CoA messages in via IPC sockets. Everything the BNG produces(events, session snapshots) gets pushed to Redis Streams, where the bng-ingestor picks them up, processes them, and persists them.Simulation and meta-configsI am generating traffic through a simulator node that mounts the host's docker socket and runs docker exec commands on selected hosts. The topology.yaml used by Containerlab to define the network topology grows bigger as more BNG's and access nodes are added. So aether.config.yaml, a simpler configuration, is consumed by the configuration pipeline to generate the topology.yaml and other files (nginx.conf, kea-dhcp.conf, RADIUS clients.conf etc.)Known Limitations- Multiple veth hops through the emulated topology add significant overhead. Profiling with iperf3 (-P 10 -t 10, 9500 MTU, 24 vCPUs) shows BNG→upstream at ~24 Gbit/s, but host→BNG→upstream drops to ~3.5 Gbit/s. The 9500 MTU also isn't representative of real ISP deployments. This gets worse when the actual network is reintroduced capping my throughput to 1.6 Gbits/sec in local.
- The circuit ID format (1/0/X) is non-standard. I simplified it for clarity.
- No iBGP or VLAN support.
- No Ipv6 support. I wanted to target IPv4 networks from the start to avoid getting too much breadth without a lot of depth.Nearly everything I know about networking (except some sections from AWS) I learned building this. A lot was figured out on the fly, so engineers will likely spot questionable decisions in the codebase. I'd genuinely appreciate that feedback.Questions- Currently, the circuit where the user connects is arbitrarily decided by the demo user. In a real system with thousands of circuits, it'd be very difficult to properly assess which circuit the customer might connect to. When adding a new customer to a service, how does the operator decide, based on customer's location, which circuit to provide the service to ?
Community Voice & Feedback
Related Early-Stage Discoveries
Discovery Source
Hacker News Aggregated via automated community intelligence tracking.
Tech Stack Dependencies
No direct open-source NPM package mentions detected in the product documentation.
Media Tractions & Mentions
No mainstream media stories specifically mentioning this product name have been intercepted yet.
Deep Research & Science
No direct peer-reviewed scientific literature matched with this product's architecture.
Market Trends