Papers
Topics
Authors
Recent
Search
2000 character limit reached

Nearest Neighbor Complexity and Boolean Circuits

Published 9 Feb 2024 in cs.CC | (2402.06740v2)

Abstract: A nearest neighbor representation of a Boolean function $f$ is a set of vectors (anchors) labeled by $0$ or $1$ such that $f(\vec{x}) = 1$ if and only if the closest anchor to $\vec{x}$ is labeled by $1$. This model was introduced by Hajnal, Liu, and Tur\'an (2022), who studied bounds on the number of anchors required to represent Boolean functions under different choices of anchors (real vs. Boolean vectors) as well as the more expressive model of $k$-nearest neighbors. We initiate the study of the representational power of nearest and $k$-nearest neighbors through Boolean circuit complexity. To this end, we establish a connection between Boolean functions with polynomial nearest neighbor complexity and those that can be efficiently represented by classes based on linear inequalities -- min-plus polynomial threshold functions -- previously studied in relation to threshold circuits. This extends an observation of Hajnal et al. (2022). We obtain exponential lower bounds on the $k$-nearest neighbors complexity of explicit $n$-variate functions, assuming $k \leq n{1-\epsilon}$. Previously, no superlinear lower bound was known for any $k>1$. Next, we further extend the connection between nearest neighbor representations and circuits to the $k$-nearest neighbors case. As a result, we show that proving superpolynomial lower bounds for the $k$-nearest neighbors complexity of an explicit function for arbitrary $k$ would require a breakthrough in circuit complexity. In addition, we prove an exponential separation between the nearest neighbor and $k$-nearest neighbors complexity (for unrestricted $k$) of an explicit function. These results address questions raised by Hajnal et al. (2022) of proving strong lower bounds for $k$-nearest neighbors and understanding the role of the parameter $k$. Finally, we devise new bounds on the nearest neighbor complexity for several explicit functions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (51)
  1. Alexandr Andoni. Nearest neighbor search: the old, the new, and the impossible. PhD thesis, Massachusetts Institute of Technology, 2009.
  2. Nearest neighbors in high-dimensional spaces. In Handbook of Discrete and Computational Geometry, pages 1135–1155. Chapman and Hall/CRC, 2017.
  3. On ACC. Comput. Complex., 4:350–366, 1994. doi: 10.1007/BF01263423. URL https://doi.org/10.1007/BF01263423.
  4. Training a 3-node neural network is NP-complete. Advances in neural information processing systems, 1, 1988.
  5. On computation and communication with small bias. In Twenty-Second Annual IEEE Conference on Computational Complexity (CCC’07), pages 24–32. IEEE, 2007.
  6. Lower bounds for linear decision lists. Chic. J. Theor. Comput. Sci., 2020, 2020. URL http://cjtcs.cs.uchicago.edu/articles/2020/1/contents.html.
  7. Rates of convergence for nearest neighbor classification. Advances in Neural Information Processing Systems, 27, 2014.
  8. Implementing the k-nearest neighbour rule via a neural network. In Proceedings of ICNN’95-International Conference on Neural Networks, volume 1, pages 136–140. IEEE, 1995.
  9. Kenneth L Clarkson. Nearest neighbor queries in metric spaces. In Proceedings of the twenty-ninth annual ACM symposium on Theory of computing, pages 609–617, 1997.
  10. Nearest neighbor pattern classification. IEEE transactions on information theory, 13(1):21–27, 1967.
  11. George Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems, 2(4):303–314, 1989.
  12. Linear threshold functions in decision lists, decision trees, and depth-2 circuits. Inf. Process. Lett., 183:106418, 2024. doi: 10.1016/J.IPL.2023.106418. URL https://doi.org/10.1016/j.ipl.2023.106418.
  13. Toward deeper understanding of neural networks: The power of initialization and a dual view on expressivity. Advances in neural information processing systems, 29, 2016.
  14. Nearest neighbor classification and search, 2020.
  15. Luc Devroye. On the asymptotic probability of error in nonparametric discrimination. The Annals of Statistics, 9(6):1320–1327, 1981.
  16. A Probabilistic Theory of Pattern Recognition, volume 31. Springer Science & Business Media, 2013.
  17. The power of depth for feedforward neural networks. In Conference on learning theory, pages 907–940. PMLR, 2016.
  18. Jürgen Forster. A linear lower bound on the unbounded error probabilistic communication complexity. Journal of Computer and System Sciences, 65(4):612–625, 2002.
  19. Christoph Globig. Case-based representability of classes of boolean functions. In Proceedings of ECAI’96, pages 117–121, 1996.
  20. Simulating threshold circuits by majority circuits. SIAM Journal on Computing, 27(1):230–246, 1998.
  21. Majority gates vs. general weighted threshold gates. Computational Complexity, 2:277–300, 1992.
  22. Threshold circuits of bounded depth. Journal of Computer and System Sciences, 46(2):129–154, 1993.
  23. Nearest neighbor representations of boolean functions. Information and Computation, 285:104879, 2022.
  24. Polynomial threshold functions and boolean threshold circuits. Information and Computation, 240:56–73, 2015.
  25. On PAC learning algorithms for rich boolean function classes. Theoretical Computer Science, 384(1):66–76, 2007.
  26. Multilayer feedforward networks are universal approximators. Neural networks, 2(5):359–366, 1989.
  27. Approximate nearest neighbors: towards removing the curse of dimensionality. In Proceedings of the thirtieth annual ACM symposium on Theory of computing, pages 604–613, 1998.
  28. Approximate nearest neighbors in limited space. In Conference On Learning Theory, pages 2012–2036. PMLR, 2018.
  29. Learnability beyond AC0. In Proceedings of the thiry-fourth annual ACM symposium on Theory of computing, pages 776–784, 2002.
  30. Stasys Jukna. Boolean function complexity: advances and frontiers, volume 5. Springer, 2012.
  31. On the information capacity of nearest neighbor representations. In 2023 IEEE International Symposium on Information Theory (ISIT), pages 1663–1668, 2023.
  32. Learning DNF in time 2O⁢(n1/3)superscript2𝑂superscript𝑛132^{O(n^{1/3})}2 start_POSTSUPERSCRIPT italic_O ( italic_n start_POSTSUPERSCRIPT 1 / 3 end_POSTSUPERSCRIPT ) end_POSTSUPERSCRIPT. Journal of Computer and System Sciences, 2(68):303–318, 2004.
  33. Toward attribute efficient learning of decision lists and parities. Journal of Machine Learning Research, 7(4), 2006.
  34. Efficient search for approximate nearest neighbor in high dimensional spaces. In Proceedings of the thirtieth annual ACM symposium on Theory of computing, pages 614–623, 1998.
  35. Constant depth circuits, fourier transform, and learnability. Journal of the ACM (JACM), 40(3):607–620, 1993.
  36. On the representational efficiency of restricted boltzmann machines. Advances in Neural Information Processing Systems, 26, 2013.
  37. O Murphy. Nearest neighbor pattern classification perceptrons. Neural Networks: Theoretical Foundations and Analysis, pages 263–266, 1992.
  38. A generalized k-nearest neighbor rule. Information and control, 16(2):128–152, 1970.
  39. Alexander A Razborov. On the distributional complexity of disjointness. In International Colloquium on Automata, Languages, and Programming, pages 249–253. Springer, 1990.
  40. Alexander A Razborov. On small depth threshold circuits. In Scandinavian Workshop on Algorithm Theory, pages 42–52. Springer, 1992.
  41. The sign-rank of AC0. SIAM Journal on Computing, 39(5):1833–1855, 2010.
  42. Ronald L Rivest. Learning decision lists. Machine learning, 2:229–246, 1987.
  43. Michael E. Saks. Slicing the hypercube, page 211–256. London Mathematical Society Lecture Note Series. Cambridge University Press, 1993.
  44. Best-case results for nearest-neighbor learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(6):599–608, 1995.
  45. Ken Satoh. Analysis of case-based representability of boolean functions by monotone theory. In Algorithmic Learning Theory: 9th International Conference, ALT’98 Otzenhausen, Germany, October 8–10, 1998 Proceedings 9, pages 179–190. Springer, 1998.
  46. Nearest-neighbor methods in learning and vision: theory and practice, volume 3. MIT press Cambridge, MA, USA:, 2005.
  47. Matus Telgarsky. Benefits of depth in neural networks. In Conference on learning theory, pages 1517–1539. PMLR, 2016.
  48. Size and depth separation in approximating benign functions with neural networks. In Conference on Learning Theory, pages 4195–4223. PMLR, 2021.
  49. Lower bounds against sparse symmetric functions of ACC circuits: Expanding the reach of #SAT algorithms. Theory Comput. Syst., 67(1):149–177, 2023. doi: 10.1007/S00224-022-10106-8. URL https://doi.org/10.1007/s00224-022-10106-8.
  50. Gordon Wilfong. Nearest neighbor problems. In Proceedings of the seventh annual symposium on Computational Geometry, pages 224–233, 1991.
  51. Richard Ryan Williams. Limits on representing boolean functions by linear combinations of simple functions: Thresholds, relus, and low-degree polynomials. In 33rd Computational Complexity Conference (CCC 2018). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, 2018.
Citations (2)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.