PocketLLM: Ultimate Compression of Large Language Models via Meta Networks