Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I've always wondered about this. Compilers do a LOT of irreversible stuff. For example, symbol names usually aren't needed (unless you have a reflective language).

Where AI would really shine is reversing the (only seemingly reversible) optimizations. For example, GCC converts "x * 14" into "(x << 4) - x - x". Of course, you can never be 100% sure the programmer didn't actually want "shift left by four followed by two subtractions", but I'm convinced that 99% of the code I write is fairly predictable and statistically similar to whatever giant codebase you train it on.



Symbol names could be inferred from context




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: