This work proposes a novel Q-learning algorithm to solve the problem of non-zero sum Nash games of linear time invariant systems with g" data-mathURL="/science?_ob=MathURL&_method=retrieve&_eid=1-s2.0-S000510981500343X&_mathId=si2.gif&_user=111111111&_pii=S000510981500343X&_rdoc=1&_issn=00051098&md5=b0c17363ebb72bae9a49f3a487882934" title="Click to view the MathML source">N-players (control inputs) and centralized uncertain/unknown dynamics. We first formulate the Q-function of each player as a parametrization of the state and all other the control inputs or players. An integral reinforcement learning approach is used to develop a model-free structure of g" data-mathURL="/science?_ob=MathURL&_method=retrieve&_eid=1-s2.0-S000510981500343X&_mathId=si2.gif&_user=111111111&_pii=S000510981500343X&_rdoc=1&_issn=00051098&md5=b0c17363ebb72bae9a49f3a487882934" title="Click to view the MathML source">N-actors/g" data-mathURL="/science?_ob=MathURL&_method=retrieve&_eid=1-s2.0-S000510981500343X&_mathId=si2.gif&_user=111111111&_pii=S000510981500343X&_rdoc=1&_issn=00051098&md5=b0c17363ebb72bae9a49f3a487882934" title="Click to view the MathML source">N-critics to estimate the parameters of the g" data-mathURL="/science?_ob=MathURL&_method=retrieve&_eid=1-s2.0-S000510981500343X&_mathId=si2.gif&_user=111111111&_pii=S000510981500343X&_rdoc=1&_issn=00051098&md5=b0c17363ebb72bae9a49f3a487882934" title="Click to view the MathML source">N-coupled Q-functions online while also guaranteeing closed-loop stability and convergence of the control policies to a Nash equilibrium. A 4th order, simulation example with five players is presented to show the efficacy of the proposed approach.