Tuesday, April 25, 2023

Relu functions

Despite their practical success, Relu activations are theoretically ill-defined. This is due to the existence of numerous trivial local optima where the gradient is zero, and we are optimizing them with approximate local optimizers.

Saturday, April 15, 2023

Optimization

#Optimization is the core of #machinelearning. All GPU acceleration, cloud systems, and transfer learning in ML are for better optimization. There are many attempts to make it faster. However, there are no widely adopted second-order methods in the field despite some momentum methods depicting them. I believe there is still potential in this area. Having a robust 10% better algorithm could have billion of dollars of impact on the industry. Of course, it is very hard to achieve.

Gemini

Dedim ki Gemini'ye yapayliktan dem vur, birkac kucuk edit yaptim ama basarili genel olarak. Ekranın ardında buz tutmuş yüzler, Sesler...