1169 lines
		
	
	
		
			48 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			1169 lines
		
	
	
		
			48 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| #
 | |
| # The Qubes OS Project, https://www.qubes-os.org/
 | |
| #
 | |
| # Copyright (C) 2014-2015
 | |
| #                   Marek Marczykowski-Górecki <marmarek@invisiblethingslab.com>
 | |
| # Copyright (C) 2015  Wojtek Porczyk <woju@invisiblethingslab.com>
 | |
| #
 | |
| # This library is free software; you can redistribute it and/or
 | |
| # modify it under the terms of the GNU Lesser General Public
 | |
| # License as published by the Free Software Foundation; either
 | |
| # version 2.1 of the License, or (at your option) any later version.
 | |
| #
 | |
| # This library is distributed in the hope that it will be useful,
 | |
| # but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| # Lesser General Public License for more details.
 | |
| #
 | |
| # You should have received a copy of the GNU Lesser General Public
 | |
| # License along with this library; if not, see <https://www.gnu.org/licenses/>.
 | |
| #
 | |
| 
 | |
| import asyncio
 | |
| import multiprocessing
 | |
| import os
 | |
| import subprocess
 | |
| import sys
 | |
| import tempfile
 | |
| import unittest
 | |
| 
 | |
| from distutils import spawn
 | |
| 
 | |
| import grp
 | |
| 
 | |
| import qubes.config
 | |
| import qubes.devices
 | |
| import qubes.tests
 | |
| import qubes.vm.appvm
 | |
| import qubes.vm.templatevm
 | |
| 
 | |
| TEST_DATA = b"0123456789" * 1024
 | |
| 
 | |
| 
 | |
| class TC_00_AppVMMixin(object):
 | |
|     def setUp(self):
 | |
|         super(TC_00_AppVMMixin, self).setUp()
 | |
|         self.init_default_template(self.template)
 | |
|         if self._testMethodName == 'test_210_time_sync':
 | |
|             self.init_networking()
 | |
|         self.testvm1 = self.app.add_new_vm(
 | |
|             qubes.vm.appvm.AppVM,
 | |
|             label='red',
 | |
|             name=self.make_vm_name('vm1'),
 | |
|             template=self.app.domains[self.template])
 | |
|         self.loop.run_until_complete(self.testvm1.create_on_disk())
 | |
|         self.testvm2 = self.app.add_new_vm(
 | |
|             qubes.vm.appvm.AppVM,
 | |
|             label='red',
 | |
|             name=self.make_vm_name('vm2'),
 | |
|             template=self.app.domains[self.template])
 | |
|         self.loop.run_until_complete(self.testvm2.create_on_disk())
 | |
|         self.app.save()
 | |
| 
 | |
|     def test_000_start_shutdown(self):
 | |
|         # TODO: wait_for, timeout
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.assertEqual(self.testvm1.get_power_state(), "Running")
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
|         self.loop.run_until_complete(self.testvm1.shutdown(wait=True))
 | |
|         self.assertEqual(self.testvm1.get_power_state(), "Halted")
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('xdotool'),
 | |
|                          "xdotool not installed")
 | |
|     def test_010_run_xterm(self):
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.assertEqual(self.testvm1.get_power_state(), "Running")
 | |
| 
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
|         p = self.loop.run_until_complete(self.testvm1.run('xterm'))
 | |
|         try:
 | |
|             title = 'user@{}'.format(self.testvm1.name)
 | |
|             if self.template.count("whonix"):
 | |
|                 title = 'user@host'
 | |
|             self.wait_for_window(title)
 | |
| 
 | |
|             self.loop.run_until_complete(asyncio.sleep(0.5))
 | |
|             subprocess.check_call(
 | |
|                 ['xdotool', 'search', '--name', title,
 | |
|                 'windowactivate', 'type', 'exit\n'])
 | |
| 
 | |
|             self.wait_for_window(title, show=False)
 | |
|         finally:
 | |
|             try:
 | |
|                 p.terminate()
 | |
|                 self.loop.run_until_complete(p.wait())
 | |
|             except ProcessLookupError:  # already dead
 | |
|                 pass
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('xdotool'),
 | |
|                          "xdotool not installed")
 | |
|     def test_011_run_gnome_terminal(self):
 | |
|         if "minimal" in self.template:
 | |
|             self.skipTest("Minimal template doesn't have 'gnome-terminal'")
 | |
|         if 'whonix' in self.template:
 | |
|             self.skipTest("Whonix template doesn't have 'gnome-terminal'")
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.assertEqual(self.testvm1.get_power_state(), "Running")
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
|         p = self.loop.run_until_complete(self.testvm1.run('gnome-terminal'))
 | |
|         try:
 | |
|             title = 'user@{}'.format(self.testvm1.name)
 | |
|             if self.template.count("whonix"):
 | |
|                 title = 'user@host'
 | |
|             self.wait_for_window(title)
 | |
| 
 | |
|             self.loop.run_until_complete(asyncio.sleep(0.5))
 | |
|             subprocess.check_call(
 | |
|                 ['xdotool', 'search', '--name', title,
 | |
|                 'windowactivate', '--sync', 'type', 'exit\n'])
 | |
| 
 | |
|             wait_count = 0
 | |
|             while subprocess.call(['xdotool', 'search', '--name', title],
 | |
|                                 stdout=open(os.path.devnull, 'w'),
 | |
|                                 stderr=subprocess.STDOUT) == 0:
 | |
|                 wait_count += 1
 | |
|                 if wait_count > 100:
 | |
|                     self.fail("Timeout while waiting for gnome-terminal "
 | |
|                             "termination")
 | |
|                 self.loop.run_until_complete(asyncio.sleep(0.1))
 | |
|         finally:
 | |
|             try:
 | |
|                 p.terminate()
 | |
|                 self.loop.run_until_complete(p.wait())
 | |
|             except ProcessLookupError:  # already dead
 | |
|                 pass
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('xdotool'),
 | |
|                          "xdotool not installed")
 | |
|     def test_012_qubes_desktop_run(self):
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.assertEqual(self.testvm1.get_power_state(), "Running")
 | |
|         xterm_desktop_path = "/usr/share/applications/xterm.desktop"
 | |
|         # Debian has it different...
 | |
|         xterm_desktop_path_debian = \
 | |
|             "/usr/share/applications/debian-xterm.desktop"
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|                 'test -r {}'.format(xterm_desktop_path_debian)))
 | |
|         except subprocess.CalledProcessError:
 | |
|             pass
 | |
|         else:
 | |
|             xterm_desktop_path = xterm_desktop_path_debian
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.run('qubes-desktop-run {}'.format(xterm_desktop_path)))
 | |
|         title = 'user@{}'.format(self.testvm1.name)
 | |
|         if self.template.count("whonix"):
 | |
|             title = 'user@host'
 | |
|         self.wait_for_window(title)
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.sleep(0.5))
 | |
|         subprocess.check_call(
 | |
|             ['xdotool', 'search', '--name', title,
 | |
|              'windowactivate', '--sync', 'type', 'exit\n'])
 | |
| 
 | |
|         self.wait_for_window(title, show=False)
 | |
| 
 | |
|     def test_050_qrexec_simple_eof(self):
 | |
|         """Test for data and EOF transmission dom0->VM"""
 | |
| 
 | |
|         # XXX is this still correct? this is no longer simple qrexec,
 | |
|         # but qubes.VMShell
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         try:
 | |
|             (stdout, stderr) = self.loop.run_until_complete(asyncio.wait_for(
 | |
|                 self.testvm1.run_for_stdio('cat', input=TEST_DATA),
 | |
|                 timeout=10))
 | |
|         except asyncio.TimeoutError:
 | |
|             self.fail(
 | |
|                 "Timeout, probably EOF wasn't transferred to the VM process")
 | |
| 
 | |
|         self.assertEqual(stdout, TEST_DATA,
 | |
|             'Received data differs from what was sent')
 | |
|         self.assertFalse(stderr,
 | |
|             'Some data was printed to stderr')
 | |
| 
 | |
|     def test_051_qrexec_simple_eof_reverse(self):
 | |
|         """Test for EOF transmission VM->dom0"""
 | |
| 
 | |
|         @asyncio.coroutine
 | |
|         def run(self):
 | |
|             p = yield from self.testvm1.run(
 | |
|                     'echo test; exec >&-; cat > /dev/null',
 | |
|                     stdin=subprocess.PIPE,
 | |
|                     stdout=subprocess.PIPE,
 | |
|                     stderr=subprocess.PIPE)
 | |
| 
 | |
|             # this will hang on test failure
 | |
|             stdout = yield from asyncio.wait_for(p.stdout.read(), timeout=10)
 | |
| 
 | |
|             p.stdin.write(TEST_DATA)
 | |
|             yield from p.stdin.drain()
 | |
|             p.stdin.close()
 | |
|             self.assertEqual(stdout.strip(), b'test',
 | |
|                 'Received data differs from what was expected')
 | |
|             # this may hang in some buggy cases
 | |
|             self.assertFalse((yield from p.stderr.read()),
 | |
|                 'Some data was printed to stderr')
 | |
| 
 | |
|             try:
 | |
|                 yield from asyncio.wait_for(p.wait(), timeout=1)
 | |
|             except asyncio.TimeoutError:
 | |
|                 self.fail("Timeout, "
 | |
|                     "probably EOF wasn't transferred from the VM process")
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
|         self.loop.run_until_complete(run(self))
 | |
| 
 | |
|     def test_052_qrexec_vm_service_eof(self):
 | |
|         """Test for EOF transmission VM(src)->VM(dst)"""
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.wait_for_session(self.testvm1),
 | |
|             self.wait_for_session(self.testvm2)]))
 | |
|         self.loop.run_until_complete(self.testvm2.run_for_stdio(
 | |
|             'cat > /etc/qubes-rpc/test.EOF',
 | |
|             user='root',
 | |
|             input=b'/bin/cat'))
 | |
| 
 | |
|         with self.qrexec_policy('test.EOF', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 stdout, _ = self.loop.run_until_complete(asyncio.wait_for(
 | |
|                     self.testvm1.run_for_stdio('''\
 | |
|                         /usr/lib/qubes/qrexec-client-vm {} test.EOF \
 | |
|                             /bin/sh -c 'echo test; exec >&-; cat >&$SAVED_FD_1'
 | |
|                     '''.format(self.testvm2.name)),
 | |
|                     timeout=10))
 | |
|             except asyncio.TimeoutError:
 | |
|                 self.fail("Timeout, probably EOF wasn't transferred")
 | |
| 
 | |
|         self.assertEqual(stdout, b'test\n',
 | |
|             'Received data differs from what was expected')
 | |
| 
 | |
|     @unittest.expectedFailure
 | |
|     def test_053_qrexec_vm_service_eof_reverse(self):
 | |
|         """Test for EOF transmission VM(src)<-VM(dst)"""
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.EOF',
 | |
|                 'echo test; exec >&-; cat >/dev/null')
 | |
| 
 | |
|         with self.qrexec_policy('test.EOF', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 stdout, _ = self.loop.run_until_complete(asyncio.wait_for(
 | |
|                     self.testvm1.run_for_stdio('''\
 | |
|                         /usr/lib/qubes/qrexec-client-vm {} test.EOF \
 | |
|                             /bin/sh -c 'cat >&$SAVED_FD_1'
 | |
|                         '''.format(self.testvm2.name)),
 | |
|                     timeout=10))
 | |
|             except asyncio.TimeoutError:
 | |
|                 self.fail("Timeout, probably EOF wasn't transferred")
 | |
| 
 | |
|         self.assertEqual(stdout, b'test',
 | |
|             'Received data differs from what was expected')
 | |
| 
 | |
|     def test_055_qrexec_dom0_service_abort(self):
 | |
|         """
 | |
|         Test if service abort (by dom0) is properly handled by source VM.
 | |
| 
 | |
|         If "remote" part of the service terminates, the source part should
 | |
|         properly be notified. This includes closing its stdin (which is
 | |
|         already checked by test_053_qrexec_vm_service_eof_reverse), but also
 | |
|         its stdout - otherwise such service might hang on write(2) call.
 | |
|         """
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.create_local_file('/etc/qubes-rpc/test.Abort',
 | |
|             'sleep 1')
 | |
| 
 | |
|         with self.qrexec_policy('test.Abort', self.testvm1, 'dom0'):
 | |
|             try:
 | |
|                 stdout, _ = self.loop.run_until_complete(asyncio.wait_for(
 | |
|                     self.testvm1.run_for_stdio('''\
 | |
|                         /usr/lib/qubes/qrexec-client-vm dom0 test.Abort \
 | |
|                             /bin/cat /dev/zero; test $? -eq 141'''),
 | |
|                     timeout=10))
 | |
|             except asyncio.TimeoutError:
 | |
|                 self.fail("Timeout, probably stdout wasn't closed")
 | |
| 
 | |
|     def test_060_qrexec_exit_code_dom0(self):
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         self.loop.run_until_complete(self.testvm1.run_for_stdio('exit 0'))
 | |
|         with self.assertRaises(subprocess.CalledProcessError) as e:
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio('exit 3'))
 | |
|         self.assertEqual(e.exception.returncode, 3)
 | |
| 
 | |
|     def test_065_qrexec_exit_code_vm(self):
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         with self.qrexec_policy('test.Retcode', self.testvm1, self.testvm2):
 | |
|             self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Retcode',
 | |
|                 'exit 0')
 | |
|             (stdout, stderr) = self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('''\
 | |
|                     /usr/lib/qubes/qrexec-client-vm {} test.Retcode;
 | |
|                         echo $?'''.format(self.testvm2.name)))
 | |
|             self.assertEqual(stdout, b'0\n')
 | |
| 
 | |
|             self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Retcode',
 | |
|                 'exit 3')
 | |
|             (stdout, stderr) = self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('''\
 | |
|                     /usr/lib/qubes/qrexec-client-vm {} test.Retcode;
 | |
|                         echo $?'''.format(self.testvm2.name)))
 | |
|             self.assertEqual(stdout, b'3\n')
 | |
| 
 | |
|     def test_070_qrexec_vm_simultaneous_write(self):
 | |
|         """Test for simultaneous write in VM(src)->VM(dst) connection
 | |
| 
 | |
|             This is regression test for #1347
 | |
| 
 | |
|             Check for deadlock when initially both sides writes a lot of data
 | |
|             (and not read anything). When one side starts reading, it should
 | |
|             get the data and the remote side should be possible to write then more.
 | |
|             There was a bug where remote side was waiting on write(2) and not
 | |
|             handling anything else.
 | |
|         """
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.write', '''\
 | |
|             # first write a lot of data
 | |
|             dd if=/dev/zero bs=993 count=10000 iflag=fullblock
 | |
|             # and only then read something
 | |
|             dd of=/dev/null bs=993 count=10000 iflag=fullblock
 | |
|             ''')
 | |
| 
 | |
|         with self.qrexec_policy('test.write', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 self.loop.run_until_complete(asyncio.wait_for(
 | |
|                     # first write a lot of data to fill all the buffers
 | |
|                     # then after some time start reading
 | |
|                     self.testvm1.run_for_stdio('''\
 | |
|                         /usr/lib/qubes/qrexec-client-vm {} test.write \
 | |
|                                 /bin/sh -c '
 | |
|                             dd if=/dev/zero bs=993 count=10000 iflag=fullblock &
 | |
|                             sleep 1;
 | |
|                             dd of=/dev/null bs=993 count=10000 iflag=fullblock;
 | |
|                             wait'
 | |
|                         '''.format(self.testvm2.name)), timeout=10))
 | |
|             except subprocess.CalledProcessError:
 | |
|                 self.fail('Service call failed')
 | |
|             except asyncio.TimeoutError:
 | |
|                 self.fail('Timeout, probably deadlock')
 | |
| 
 | |
|     def test_071_qrexec_dom0_simultaneous_write(self):
 | |
|         """Test for simultaneous write in dom0(src)->VM(dst) connection
 | |
| 
 | |
|             Similar to test_070_qrexec_vm_simultaneous_write, but with dom0
 | |
|             as a source.
 | |
|         """
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm2.start())
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.write', '''\
 | |
|             # first write a lot of data
 | |
|             dd if=/dev/zero bs=993 count=10000 iflag=fullblock
 | |
|             # and only then read something
 | |
|             dd of=/dev/null bs=993 count=10000 iflag=fullblock
 | |
|             ''')
 | |
| 
 | |
|         # can't use subprocess.PIPE, because asyncio will claim those FDs
 | |
|         pipe1_r, pipe1_w = os.pipe()
 | |
|         pipe2_r, pipe2_w = os.pipe()
 | |
|         try:
 | |
|             local_proc = self.loop.run_until_complete(
 | |
|                 asyncio.create_subprocess_shell(
 | |
|                     # first write a lot of data to fill all the buffers
 | |
|                     "dd if=/dev/zero bs=993 count=10000 iflag=fullblock & "
 | |
|                     # then after some time start reading
 | |
|                     "sleep 1; "
 | |
|                     "dd of=/dev/null bs=993 count=10000 iflag=fullblock; "
 | |
|                     "wait", stdin=pipe1_r, stdout=pipe2_w))
 | |
| 
 | |
|             service_proc = self.loop.run_until_complete(self.testvm2.run_service(
 | |
|                 "test.write", stdin=pipe2_r, stdout=pipe1_w))
 | |
|         finally:
 | |
|             os.close(pipe1_r)
 | |
|             os.close(pipe1_w)
 | |
|             os.close(pipe2_r)
 | |
|             os.close(pipe2_w)
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(
 | |
|                 asyncio.wait_for(service_proc.wait(), timeout=10))
 | |
|         except asyncio.TimeoutError:
 | |
|             self.fail("Timeout, probably deadlock")
 | |
|         else:
 | |
|             self.assertEqual(service_proc.returncode, 0,
 | |
|                 "Service call failed")
 | |
|         finally:
 | |
|             try:
 | |
|                 service_proc.terminate()
 | |
|             except ProcessLookupError:
 | |
|                 pass
 | |
| 
 | |
|     def test_072_qrexec_to_dom0_simultaneous_write(self):
 | |
|         """Test for simultaneous write in dom0(src)<-VM(dst) connection
 | |
| 
 | |
|             Similar to test_071_qrexec_dom0_simultaneous_write, but with dom0
 | |
|             as a "hanging" side.
 | |
|         """
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm2.start())
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.write', '''\
 | |
|             # first write a lot of data
 | |
|             dd if=/dev/zero bs=993 count=10000 iflag=fullblock &
 | |
|             # and only then read something
 | |
|             dd of=/dev/null bs=993 count=10000 iflag=fullblock
 | |
|             sleep 1;
 | |
|             wait
 | |
|             ''')
 | |
| 
 | |
|         # can't use subprocess.PIPE, because asyncio will claim those FDs
 | |
|         pipe1_r, pipe1_w = os.pipe()
 | |
|         pipe2_r, pipe2_w = os.pipe()
 | |
|         try:
 | |
|             local_proc = self.loop.run_until_complete(
 | |
|                 asyncio.create_subprocess_shell(
 | |
|                     # first write a lot of data to fill all the buffers
 | |
|                     "dd if=/dev/zero bs=993 count=10000 iflag=fullblock & "
 | |
|                     # then, only when all written, read something
 | |
|                     "dd of=/dev/null bs=993 count=10000 iflag=fullblock; ",
 | |
|                     stdin=pipe1_r, stdout=pipe2_w))
 | |
| 
 | |
|             service_proc = self.loop.run_until_complete(self.testvm2.run_service(
 | |
|                 "test.write", stdin=pipe2_r, stdout=pipe1_w))
 | |
|         finally:
 | |
|             os.close(pipe1_r)
 | |
|             os.close(pipe1_w)
 | |
|             os.close(pipe2_r)
 | |
|             os.close(pipe2_w)
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(
 | |
|                 asyncio.wait_for(service_proc.wait(), timeout=10))
 | |
|         except asyncio.TimeoutError:
 | |
|             self.fail("Timeout, probably deadlock")
 | |
|         else:
 | |
|             self.assertEqual(service_proc.returncode, 0,
 | |
|                 "Service call failed")
 | |
|         finally:
 | |
|             try:
 | |
|                 service_proc.terminate()
 | |
|             except ProcessLookupError:
 | |
|                 pass
 | |
| 
 | |
|     def test_080_qrexec_service_argument_allow_default(self):
 | |
|         """Qrexec service call with argument"""
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Argument',
 | |
|             '/usr/bin/printf %s "$1"')
 | |
|         with self.qrexec_policy('test.Argument', self.testvm1, self.testvm2):
 | |
|             stdout, stderr = self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('/usr/lib/qubes/qrexec-client-vm '
 | |
|                     '{} test.Argument+argument'.format(self.testvm2.name)))
 | |
|             self.assertEqual(stdout, b'argument')
 | |
| 
 | |
|     def test_081_qrexec_service_argument_allow_specific(self):
 | |
|         """Qrexec service call with argument - allow only specific value"""
 | |
| 
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Argument',
 | |
|             '/usr/bin/printf %s "$1"')
 | |
| 
 | |
|         with self.qrexec_policy('test.Argument', '$anyvm', '$anyvm', False):
 | |
|             with self.qrexec_policy('test.Argument+argument',
 | |
|                     self.testvm1.name, self.testvm2.name):
 | |
|                 stdout, stderr = self.loop.run_until_complete(
 | |
|                     self.testvm1.run_for_stdio(
 | |
|                         '/usr/lib/qubes/qrexec-client-vm '
 | |
|                         '{} test.Argument+argument'.format(self.testvm2.name)))
 | |
|         self.assertEqual(stdout, b'argument')
 | |
| 
 | |
|     def test_082_qrexec_service_argument_deny_specific(self):
 | |
|         """Qrexec service call with argument - deny specific value"""
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Argument',
 | |
|             '/usr/bin/printf %s "$1"')
 | |
|         with self.qrexec_policy('test.Argument', '$anyvm', '$anyvm'):
 | |
|             with self.qrexec_policy('test.Argument+argument',
 | |
|                     self.testvm1, self.testvm2, allow=False):
 | |
|                 with self.assertRaises(subprocess.CalledProcessError,
 | |
|                         msg='Service request should be denied'):
 | |
|                     self.loop.run_until_complete(
 | |
|                         self.testvm1.run_for_stdio(
 | |
|                             '/usr/lib/qubes/qrexec-client-vm {} '
 | |
|                             'test.Argument+argument'.format(self.testvm2.name)))
 | |
| 
 | |
|     def test_083_qrexec_service_argument_specific_implementation(self):
 | |
|         """Qrexec service call with argument - argument specific
 | |
|         implementatation"""
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2,
 | |
|             '/etc/qubes-rpc/test.Argument',
 | |
|             '/usr/bin/printf %s "$1"')
 | |
|         self.create_remote_file(self.testvm2,
 | |
|             '/etc/qubes-rpc/test.Argument+argument',
 | |
|             '/usr/bin/printf "specific: %s" "$1"')
 | |
| 
 | |
|         with self.qrexec_policy('test.Argument', self.testvm1, self.testvm2):
 | |
|             stdout, stderr = self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('/usr/lib/qubes/qrexec-client-vm '
 | |
|                     '{} test.Argument+argument'.format(self.testvm2.name)))
 | |
| 
 | |
|         self.assertEqual(stdout, b'specific: argument')
 | |
| 
 | |
|     def test_084_qrexec_service_argument_extra_env(self):
 | |
|         """Qrexec service call with argument - extra env variables"""
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.create_remote_file(self.testvm2, '/etc/qubes-rpc/test.Argument',
 | |
|             '/usr/bin/printf "%s %s" '
 | |
|                 '"$QREXEC_SERVICE_FULL_NAME" "$QREXEC_SERVICE_ARGUMENT"')
 | |
| 
 | |
|         with self.qrexec_policy('test.Argument', self.testvm1, self.testvm2):
 | |
|             stdout, stderr = self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('/usr/lib/qubes/qrexec-client-vm '
 | |
|                     '{} test.Argument+argument'.format(self.testvm2.name)))
 | |
| 
 | |
|         self.assertEqual(stdout, b'test.Argument+argument argument')
 | |
| 
 | |
|     def test_100_qrexec_filecopy(self):
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         with self.qrexec_policy('qubes.Filecopy', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 self.loop.run_until_complete(
 | |
|                     self.testvm1.run_for_stdio(
 | |
|                         'qvm-copy-to-vm {} /etc/passwd'.format(
 | |
|                             self.testvm2.name)))
 | |
|             except subprocess.CalledProcessError as e:
 | |
|                 self.fail('qvm-copy-to-vm failed: {}'.format(e.stderr))
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm2.run_for_stdio(
 | |
|                 'diff /etc/passwd /home/user/QubesIncoming/{}/passwd'.format(
 | |
|                     self.testvm1.name)))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.fail('file differs')
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|                 'test -f /etc/passwd'))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.fail('source file got removed')
 | |
| 
 | |
|     def test_105_qrexec_filemove(self):
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|             'cp /etc/passwd /tmp/passwd'))
 | |
|         with self.qrexec_policy('qubes.Filecopy', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 self.loop.run_until_complete(
 | |
|                     self.testvm1.run_for_stdio(
 | |
|                         'qvm-move-to-vm {} /tmp/passwd'.format(
 | |
|                             self.testvm2.name)))
 | |
|             except subprocess.CalledProcessError as e:
 | |
|                 self.fail('qvm-move-to-vm failed: {}'.format(e.stderr))
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm2.run_for_stdio(
 | |
|                 'diff /etc/passwd /home/user/QubesIncoming/{}/passwd'.format(
 | |
|                     self.testvm1.name)))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.fail('file differs')
 | |
| 
 | |
|         with self.assertRaises(subprocess.CalledProcessError):
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|                 'test -f /tmp/passwd'))
 | |
| 
 | |
|     def test_101_qrexec_filecopy_with_autostart(self):
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
| 
 | |
|         with self.qrexec_policy('qubes.Filecopy', self.testvm1, self.testvm2):
 | |
|             try:
 | |
|                 self.loop.run_until_complete(
 | |
|                     self.testvm1.run_for_stdio(
 | |
|                         'qvm-copy-to-vm {} /etc/passwd'.format(
 | |
|                             self.testvm2.name)))
 | |
|             except subprocess.CalledProcessError as e:
 | |
|                 self.fail('qvm-copy-to-vm failed: {}'.format(e.stderr))
 | |
| 
 | |
|         # workaround for libvirt bug (domain ID isn't updated when is started
 | |
|         #  from other application) - details in
 | |
|         # QubesOS/qubes-core-libvirt@63ede4dfb4485c4161dd6a2cc809e8fb45ca664f
 | |
|         # XXX is it still true with qubesd? --woju 20170523
 | |
|         self.testvm2._libvirt_domain = None
 | |
|         self.assertTrue(self.testvm2.is_running())
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm2.run_for_stdio(
 | |
|                 'diff /etc/passwd /home/user/QubesIncoming/{}/passwd'.format(
 | |
|                     self.testvm1.name)))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.fail('file differs')
 | |
| 
 | |
|         try:
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|                 'test -f /etc/passwd'))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.fail('source file got removed')
 | |
| 
 | |
|     def test_110_qrexec_filecopy_deny(self):
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         with self.qrexec_policy('qubes.Filecopy', self.testvm1, self.testvm2,
 | |
|                 allow=False):
 | |
|             with self.assertRaises(subprocess.CalledProcessError):
 | |
|                 self.loop.run_until_complete(
 | |
|                     self.testvm1.run_for_stdio(
 | |
|                         'qvm-copy-to-vm {} /etc/passwd'.format(
 | |
|                             self.testvm2.name)))
 | |
| 
 | |
|         with self.assertRaises(subprocess.CalledProcessError):
 | |
|             self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|                 'test -d /home/user/QubesIncoming/{}'.format(
 | |
|                     self.testvm1.name)))
 | |
| 
 | |
|     @unittest.skip("Xen gntalloc driver crashes when page is mapped in the "
 | |
|                    "same domain")
 | |
|     def test_120_qrexec_filecopy_self(self):
 | |
|         self.testvm1.start()
 | |
|         self.qrexec_policy('qubes.Filecopy', self.testvm1.name,
 | |
|             self.testvm1.name)
 | |
|         p = self.testvm1.run("qvm-copy-to-vm %s /etc/passwd" %
 | |
|                              self.testvm1.name, passio_popen=True,
 | |
|                              passio_stderr=True)
 | |
|         p.wait()
 | |
|         self.assertEqual(p.returncode, 0, "qvm-copy-to-vm failed: %s" %
 | |
|                          p.stderr.read())
 | |
|         retcode = self.testvm1.run(
 | |
|             "diff /etc/passwd /home/user/QubesIncoming/{}/passwd".format(
 | |
|                 self.testvm1.name),
 | |
|             wait=True)
 | |
|         self.assertEqual(retcode, 0, "file differs")
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('xdotool'),
 | |
|                          "xdotool not installed")
 | |
|     def test_130_qrexec_filemove_disk_full(self):
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start()]))
 | |
| 
 | |
|         self.loop.run_until_complete(self.wait_for_session(self.testvm1))
 | |
| 
 | |
|         # Prepare test file
 | |
|         self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|             'yes teststring | dd of=/tmp/testfile bs=1M count=50 '
 | |
|             'iflag=fullblock'))
 | |
| 
 | |
|         # Prepare target directory with limited size
 | |
|         self.loop.run_until_complete(self.testvm2.run_for_stdio(
 | |
|             'mkdir -p /home/user/QubesIncoming && '
 | |
|             'chown user /home/user/QubesIncoming && '
 | |
|             'mount -t tmpfs none /home/user/QubesIncoming -o size=48M',
 | |
|             user='root'))
 | |
| 
 | |
|         with self.qrexec_policy('qubes.Filecopy', self.testvm1, self.testvm2):
 | |
|             p = self.loop.run_until_complete(self.testvm1.run(
 | |
|                 'qvm-move-to-vm {} /tmp/testfile'.format(
 | |
|                     self.testvm2.name)))
 | |
| 
 | |
|             # Close GUI error message
 | |
|             try:
 | |
|                 self.enter_keys_in_window('Error', ['Return'])
 | |
|             except subprocess.CalledProcessError:
 | |
|                 pass
 | |
|             self.loop.run_until_complete(p.wait())
 | |
|             self.assertNotEqual(p.returncode, 0)
 | |
| 
 | |
|         # the file shouldn't be removed in source vm
 | |
|         self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|             'test -f /tmp/testfile'))
 | |
| 
 | |
|     def test_200_timezone(self):
 | |
|         """Test whether timezone setting is properly propagated to the VM"""
 | |
|         if "whonix" in self.template:
 | |
|             self.skipTest("Timezone propagation disabled on Whonix templates")
 | |
| 
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         vm_tz, _ = self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|             'date +%Z'))
 | |
|         dom0_tz = subprocess.check_output(['date', '+%Z'])
 | |
|         self.assertEqual(vm_tz.strip(), dom0_tz.strip())
 | |
| 
 | |
|         # Check if reverting back to UTC works
 | |
|         vm_tz, _ = self.loop.run_until_complete(self.testvm1.run_for_stdio(
 | |
|             'TZ=UTC date +%Z'))
 | |
|         self.assertEqual(vm_tz.strip(), b'UTC')
 | |
| 
 | |
|     def test_210_time_sync(self):
 | |
|         """Test time synchronization mechanism"""
 | |
|         if self.template.startswith('whonix-'):
 | |
|             self.skipTest('qvm-sync-clock disabled for Whonix VMs')
 | |
|         self.loop.run_until_complete(asyncio.wait([
 | |
|             self.testvm1.start(),
 | |
|             self.testvm2.start(),]))
 | |
|         start_time = subprocess.check_output(['date', '-u', '+%s'])
 | |
| 
 | |
|         try:
 | |
|             self.app.clockvm = self.testvm1
 | |
|             self.app.save()
 | |
|             # break vm and dom0 time, to check if qvm-sync-clock would fix it
 | |
|             subprocess.check_call(['sudo', 'date', '-s', '2001-01-01T12:34:56'],
 | |
|                 stdout=subprocess.DEVNULL)
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm2.run_for_stdio('date -s 2001-01-01T12:34:56',
 | |
|                     user='root'))
 | |
| 
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm2.run_for_stdio('qvm-sync-clock',
 | |
|                     user='root'))
 | |
| 
 | |
|             p = self.loop.run_until_complete(
 | |
|                 asyncio.create_subprocess_exec('sudo', 'qvm-sync-clock',
 | |
|                     stdout=asyncio.subprocess.DEVNULL))
 | |
|             self.loop.run_until_complete(p.wait())
 | |
|             self.assertEqual(p.returncode, 0)
 | |
|             vm_time, _ = self.loop.run_until_complete(
 | |
|                 self.testvm2.run_for_stdio('date -u +%s'))
 | |
|             self.assertAlmostEquals(int(vm_time), int(start_time), delta=30)
 | |
| 
 | |
|             dom0_time = subprocess.check_output(['date', '-u', '+%s'])
 | |
|             self.assertAlmostEquals(int(dom0_time), int(start_time), delta=30)
 | |
| 
 | |
|         except:
 | |
|             # reset time to some approximation of the real time
 | |
|             subprocess.Popen(
 | |
|                 ["sudo", "date", "-u", "-s", "@" + start_time.decode()])
 | |
|             raise
 | |
|         finally:
 | |
|             self.app.clockvm = None
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('parecord'),
 | |
|                          "pulseaudio-utils not installed in dom0")
 | |
|     def test_220_audio_playback(self):
 | |
|         if 'whonix-gw' in self.template:
 | |
|             self.skipTest('whonix-gw have no audio')
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         try:
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('which parecord'))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.skipTest('pulseaudio-utils not installed in VM')
 | |
| 
 | |
|         self.loop.run_until_complete(
 | |
|             self.wait_for_session(self.testvm1))
 | |
|         # and some more...
 | |
|         self.loop.run_until_complete(asyncio.sleep(1))
 | |
|         # generate some "audio" data
 | |
|         audio_in = b'\x20' * 44100
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio('cat > audio_in.raw', input=audio_in))
 | |
|         local_user = grp.getgrnam('qubes').gr_mem[0]
 | |
|         with tempfile.NamedTemporaryFile() as recorded_audio:
 | |
|             os.chmod(recorded_audio.name, 0o666)
 | |
|             # FIXME: -d 0 assumes only one audio device
 | |
|             p = subprocess.Popen(['sudo', '-E', '-u', local_user,
 | |
|                 'parecord', '-d', '0', '--raw', recorded_audio.name],
 | |
|                 stdout=subprocess.PIPE)
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('paplay --raw audio_in.raw'))
 | |
|             # wait for possible parecord buffering
 | |
|             self.loop.run_until_complete(asyncio.sleep(1))
 | |
|             p.terminate()
 | |
|             # for some reason sudo do not relay SIGTERM sent above
 | |
|             subprocess.check_call(['pkill', 'parecord'])
 | |
|             p.wait()
 | |
|             # allow few bytes missing, don't use assertIn, to avoid printing
 | |
|             # the whole data in error message
 | |
|             if audio_in[:-8] not in recorded_audio.file.read():
 | |
|                 self.fail('played sound not found in dom0')
 | |
| 
 | |
|     def _configure_audio_recording(self, vm):
 | |
|         '''Connect VM's output-source to sink monitor instead of mic'''
 | |
|         local_user = grp.getgrnam('qubes').gr_mem[0]
 | |
|         sudo = ['sudo', '-E', '-u', local_user]
 | |
|         source_outputs = subprocess.check_output(
 | |
|             sudo + ['pacmd', 'list-source-outputs']).decode()
 | |
| 
 | |
|         last_index = None
 | |
|         found = False
 | |
|         for line in source_outputs.splitlines():
 | |
|             if line.startswith('    index: '):
 | |
|                 last_index = line.split(':')[1].strip()
 | |
|             elif line.startswith('\t\tapplication.name = '):
 | |
|                 app_name = line.split('=')[1].strip('" ')
 | |
|                 if vm.name == app_name:
 | |
|                     found = True
 | |
|                     break
 | |
|         if not found:
 | |
|             self.fail('source-output for VM {} not found'.format(vm.name))
 | |
| 
 | |
|         subprocess.check_call(sudo +
 | |
|             ['pacmd', 'move-source-output', last_index, '0'])
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('parecord'),
 | |
|                          "pulseaudio-utils not installed in dom0")
 | |
|     def test_221_audio_record_muted(self):
 | |
|         if 'whonix-gw' in self.template:
 | |
|             self.skipTest('whonix-gw have no audio')
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         try:
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('which parecord'))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.skipTest('pulseaudio-utils not installed in VM')
 | |
| 
 | |
|         self.loop.run_until_complete(
 | |
|             self.wait_for_session(self.testvm1))
 | |
|         # and some more...
 | |
|         self.loop.run_until_complete(asyncio.sleep(1))
 | |
|         # connect VM's recording source output monitor (instead of mic)
 | |
|         self._configure_audio_recording(self.testvm1)
 | |
| 
 | |
|         # generate some "audio" data
 | |
|         audio_in = b'\x20' * 44100
 | |
|         local_user = grp.getgrnam('qubes').gr_mem[0]
 | |
|         record = self.loop.run_until_complete(
 | |
|             self.testvm1.run('parecord --raw audio_rec.raw'))
 | |
|         # give it time to start recording
 | |
|         self.loop.run_until_complete(asyncio.sleep(0.5))
 | |
|         p = subprocess.Popen(['sudo', '-E', '-u', local_user,
 | |
|             'paplay', '--raw'],
 | |
|             stdin=subprocess.PIPE)
 | |
|         p.communicate(audio_in)
 | |
|         # wait for possible parecord buffering
 | |
|         self.loop.run_until_complete(asyncio.sleep(1))
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio('pkill parecord'))
 | |
|         record.wait()
 | |
|         recorded_audio, _ = self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio('cat audio_rec.raw'))
 | |
|         # should be empty or silence, so check just a little fragment
 | |
|         if audio_in[:32] in recorded_audio:
 | |
|             self.fail('VM recorded something, even though mic disabled')
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('parecord'),
 | |
|                          "pulseaudio-utils not installed in dom0")
 | |
|     def test_222_audio_record_unmuted(self):
 | |
|         if 'whonix-gw' in self.template:
 | |
|             self.skipTest('whonix-gw have no audio')
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         try:
 | |
|             self.loop.run_until_complete(
 | |
|                 self.testvm1.run_for_stdio('which parecord'))
 | |
|         except subprocess.CalledProcessError:
 | |
|             self.skipTest('pulseaudio-utils not installed in VM')
 | |
| 
 | |
|         self.loop.run_until_complete(
 | |
|             self.wait_for_session(self.testvm1))
 | |
|         # and some more...
 | |
|         self.loop.run_until_complete(asyncio.sleep(1))
 | |
|         da = qubes.devices.DeviceAssignment(self.app.domains[0], 'mic')
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.devices['mic'].attach(da))
 | |
|         # connect VM's recording source output monitor (instead of mic)
 | |
|         self._configure_audio_recording(self.testvm1)
 | |
| 
 | |
|         # generate some "audio" data
 | |
|         audio_in = b'\x20' * 44100
 | |
|         local_user = grp.getgrnam('qubes').gr_mem[0]
 | |
|         record = self.loop.run_until_complete(
 | |
|             self.testvm1.run('parecord --raw audio_rec.raw'))
 | |
|         # give it time to start recording
 | |
|         self.loop.run_until_complete(asyncio.sleep(0.5))
 | |
|         p = subprocess.Popen(['sudo', '-E', '-u', local_user,
 | |
|             'paplay', '--raw'],
 | |
|             stdin=subprocess.PIPE)
 | |
|         p.communicate(audio_in)
 | |
|         # wait for possible parecord buffering
 | |
|         self.loop.run_until_complete(asyncio.sleep(1))
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio('pkill parecord'))
 | |
|         record.wait()
 | |
|         recorded_audio, _ = self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio('cat audio_rec.raw'))
 | |
|         # allow few bytes to be missing
 | |
|         if audio_in[:-8] not in recorded_audio:
 | |
|             self.fail('VM not recorded expected data')
 | |
| 
 | |
|     def test_250_resize_private_img(self):
 | |
|         """
 | |
|         Test private.img resize, both offline and online
 | |
|         :return:
 | |
|         """
 | |
|         # First offline test
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.storage.resize('private', 4*1024**3))
 | |
|         self.loop.run_until_complete(self.testvm1.start())
 | |
|         df_cmd = '( df --output=size /rw || df /rw | awk \'{print $2}\' )|' \
 | |
|                  'tail -n 1'
 | |
|         # new_size in 1k-blocks
 | |
|         new_size, _ = self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio(df_cmd))
 | |
|         # some safety margin for FS metadata
 | |
|         self.assertGreater(int(new_size.strip()), 3.8*1024**2)
 | |
|         # Then online test
 | |
|         self.loop.run_until_complete(
 | |
|             self.testvm1.storage.resize('private', 6*1024**3))
 | |
|         # new_size in 1k-blocks
 | |
|         new_size, _ = self.loop.run_until_complete(
 | |
|             self.testvm1.run_for_stdio(df_cmd))
 | |
|         # some safety margin for FS metadata
 | |
|         self.assertGreater(int(new_size.strip()), 5.7*1024**2)
 | |
| 
 | |
|     @unittest.skipUnless(spawn.find_executable('xdotool'),
 | |
|                          "xdotool not installed")
 | |
|     def test_300_bug_1028_gui_memory_pinning(self):
 | |
|         """
 | |
|         If VM window composition buffers are relocated in memory, GUI will
 | |
|         still use old pointers and will display old pages
 | |
|         :return:
 | |
|         """
 | |
| 
 | |
|         # this test does too much asynchronous operations,
 | |
|         # so let's rewrite it as a coroutine and call it as such
 | |
|         return self.loop.run_until_complete(
 | |
|             self._test_300_bug_1028_gui_memory_pinning())
 | |
| 
 | |
|     @asyncio.coroutine
 | |
|     def _test_300_bug_1028_gui_memory_pinning(self):
 | |
|         self.testvm1.memory = 800
 | |
|         self.testvm1.maxmem = 800
 | |
| 
 | |
|         # exclude from memory balancing
 | |
|         self.testvm1.features['service.meminfo-writer'] = False
 | |
|         yield from self.testvm1.start()
 | |
|         yield from self.wait_for_session(self.testvm1)
 | |
| 
 | |
|         # and allow large map count
 | |
|         yield from self.testvm1.run('echo 256000 > /proc/sys/vm/max_map_count',
 | |
|             user="root")
 | |
| 
 | |
|         allocator_c = '''
 | |
| #include <sys/mman.h>
 | |
| #include <stdlib.h>
 | |
| #include <stdio.h>
 | |
| 
 | |
| int main(int argc, char **argv) {
 | |
|     int total_pages;
 | |
|     char *addr, *iter;
 | |
| 
 | |
|     total_pages = atoi(argv[1]);
 | |
|     addr = mmap(NULL, total_pages * 0x1000, PROT_READ | PROT_WRITE,
 | |
|         MAP_ANONYMOUS | MAP_PRIVATE | MAP_POPULATE, -1, 0);
 | |
|     if (addr == MAP_FAILED) {
 | |
|         perror("mmap");
 | |
|         exit(1);
 | |
|     }
 | |
| 
 | |
|     printf("Stage1\\n");
 | |
|     fflush(stdout);
 | |
|     getchar();
 | |
|     for (iter = addr; iter < addr + total_pages*0x1000; iter += 0x2000) {
 | |
|         if (mlock(iter, 0x1000) == -1) {
 | |
|             perror("mlock");
 | |
|             fprintf(stderr, "%d of %d\\n", (iter-addr)/0x1000, total_pages);
 | |
|             exit(1);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     printf("Stage2\\n");
 | |
|     fflush(stdout);
 | |
|     for (iter = addr+0x1000; iter < addr + total_pages*0x1000; iter += 0x2000) {
 | |
|         if (munmap(iter, 0x1000) == -1) {
 | |
|             perror(\"munmap\");
 | |
|             exit(1);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     printf("Stage3\\n");
 | |
|     fflush(stdout);
 | |
|     fclose(stdout);
 | |
|     getchar();
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| '''
 | |
| 
 | |
|         yield from self.testvm1.run_for_stdio('cat > allocator.c',
 | |
|             input=allocator_c.encode())
 | |
| 
 | |
|         try:
 | |
|             yield from self.testvm1.run_for_stdio(
 | |
|                 'gcc allocator.c -o allocator')
 | |
|         except subprocess.CalledProcessError as e:
 | |
|             self.skipTest('allocator compile failed: {}'.format(e.stderr))
 | |
| 
 | |
|         # drop caches to have even more memory pressure
 | |
|         yield from self.testvm1.run_for_stdio(
 | |
|             'echo 3 > /proc/sys/vm/drop_caches', user='root')
 | |
| 
 | |
|         # now fragment all free memory
 | |
|         stdout, _ = yield from self.testvm1.run_for_stdio(
 | |
|             "grep ^MemFree: /proc/meminfo|awk '{print $2}'")
 | |
|         memory_pages = int(stdout) // 4  # 4k pages
 | |
| 
 | |
|         alloc1 = yield from self.testvm1.run(
 | |
|             'ulimit -l unlimited; exec /home/user/allocator {}'.format(
 | |
|                 memory_pages),
 | |
|             user="root",
 | |
|             stdin=subprocess.PIPE, stdout=subprocess.PIPE,
 | |
|             stderr=subprocess.PIPE)
 | |
| 
 | |
|         # wait for memory being allocated; can't use just .read(), because EOF
 | |
|         # passing is unreliable while the process is still running
 | |
|         alloc1.stdin.write(b'\n')
 | |
|         yield from alloc1.stdin.drain()
 | |
|         try:
 | |
|             alloc_out = yield from alloc1.stdout.readexactly(
 | |
|                 len('Stage1\nStage2\nStage3\n'))
 | |
|         except asyncio.IncompleteReadError as e:
 | |
|             alloc_out = e.partial
 | |
| 
 | |
|         if b'Stage3' not in alloc_out:
 | |
|             # read stderr only in case of failed assert (), but still have nice
 | |
|             # failure message (don't use self.fail() directly)
 | |
|             #
 | |
|             # stderr isn't always read, because on not-failed run, the process
 | |
|             # is still running, so stderr.read() will wait (indefinitely).
 | |
|             self.assertIn(b'Stage3', alloc_out,
 | |
|                 (yield from alloc1.stderr.read()))
 | |
| 
 | |
|         # now, launch some window - it should get fragmented composition buffer
 | |
|         # it is important to have some changing content there, to generate
 | |
|         # content update events (aka damage notify)
 | |
|         proc = yield from self.testvm1.run(
 | |
|             'xterm -maximized -e top')
 | |
| 
 | |
|         if proc.returncode is not None:
 | |
|             self.fail('xterm failed to start')
 | |
|         # get window ID
 | |
|         winid = yield from self.wait_for_window_coro(
 | |
|             self.testvm1.name + ':xterm',
 | |
|             search_class=True)
 | |
|         xprop = yield from asyncio.get_event_loop().run_in_executor(None,
 | |
|             subprocess.check_output,
 | |
|             ['xprop', '-notype', '-id', winid, '_QUBES_VMWINDOWID'])
 | |
|         vm_winid = xprop.decode().strip().split(' ')[4]
 | |
| 
 | |
|         # now free the fragmented memory and trigger compaction
 | |
|         alloc1.stdin.write(b'\n')
 | |
|         yield from alloc1.stdin.drain()
 | |
|         yield from alloc1.wait()
 | |
|         yield from self.testvm1.run_for_stdio(
 | |
|             'echo 1 > /proc/sys/vm/compact_memory', user='root')
 | |
| 
 | |
|         # now window may be already "broken"; to be sure, allocate (=zero)
 | |
|         # some memory
 | |
|         alloc2 = yield from self.testvm1.run(
 | |
|             'ulimit -l unlimited; /home/user/allocator {}'.format(memory_pages),
 | |
|             user='root', stdout=subprocess.PIPE)
 | |
|         yield from alloc2.stdout.read(len('Stage1\n'))
 | |
| 
 | |
|         # wait for damage notify - top updates every 3 sec by default
 | |
|         yield from asyncio.sleep(6)
 | |
| 
 | |
|         # stop changing the window content
 | |
|         subprocess.check_call(['xdotool', 'key', '--window', winid, 'd'])
 | |
| 
 | |
|         # now take screenshot of the window, from dom0 and VM
 | |
|         # choose pnm format, as it doesn't have any useless metadata - easy
 | |
|         # to compare
 | |
|         vm_image, _ = yield from self.testvm1.run_for_stdio(
 | |
|             'import -window {} pnm:-'.format(vm_winid))
 | |
| 
 | |
|         dom0_image = yield from asyncio.get_event_loop().run_in_executor(None,
 | |
|             subprocess.check_output, ['import', '-window', winid, 'pnm:-'])
 | |
| 
 | |
|         if vm_image != dom0_image:
 | |
|             self.fail("Dom0 window doesn't match VM window content")
 | |
| 
 | |
| class TC_10_Generic(qubes.tests.SystemTestCase):
 | |
|     def setUp(self):
 | |
|         super(TC_10_Generic, self).setUp()
 | |
|         self.init_default_template()
 | |
|         self.vm = self.app.add_new_vm(
 | |
|             qubes.vm.appvm.AppVM,
 | |
|             name=self.make_vm_name('vm'),
 | |
|             label='red',
 | |
|             template=self.app.default_template)
 | |
|         self.loop.run_until_complete(self.vm.create_on_disk())
 | |
|         self.app.save()
 | |
|         self.vm = self.app.domains[self.vm.qid]
 | |
| 
 | |
|     def test_000_anyvm_deny_dom0(self):
 | |
|         '''$anyvm in policy should not match dom0'''
 | |
|         policy = open("/etc/qubes-rpc/policy/test.AnyvmDeny", "w")
 | |
|         policy.write("%s $anyvm allow" % (self.vm.name,))
 | |
|         policy.close()
 | |
|         self.addCleanup(os.unlink, "/etc/qubes-rpc/policy/test.AnyvmDeny")
 | |
| 
 | |
|         flagfile = '/tmp/test-anyvmdeny-flag'
 | |
|         if os.path.exists(flagfile):
 | |
|             os.remove(flagfile)
 | |
| 
 | |
|         self.create_local_file('/etc/qubes-rpc/test.AnyvmDeny',
 | |
|             'touch {}\necho service output\n'.format(flagfile))
 | |
| 
 | |
|         self.loop.run_until_complete(self.vm.start())
 | |
|         with self.qrexec_policy('test.AnyvmDeny', self.vm, '$anyvm'):
 | |
|             with self.assertRaises(subprocess.CalledProcessError,
 | |
|                     msg='$anyvm matched dom0') as e:
 | |
|                 self.loop.run_until_complete(
 | |
|                     self.vm.run_for_stdio(
 | |
|                         '/usr/lib/qubes/qrexec-client-vm dom0 test.AnyvmDeny'))
 | |
|             stdout = e.exception.output
 | |
|             stderr = e.exception.stderr
 | |
|         self.assertFalse(os.path.exists(flagfile),
 | |
|             'Flag file created (service was run) even though should be denied,'
 | |
|             ' qrexec-client-vm output: {} {}'.format(stdout, stderr))
 | |
| 
 | |
| def create_testcases_for_templates():
 | |
|     return qubes.tests.create_testcases_for_templates('TC_00_AppVM',
 | |
|         TC_00_AppVMMixin, qubes.tests.SystemTestCase,
 | |
|         module=sys.modules[__name__])
 | |
| 
 | |
| def load_tests(loader, tests, pattern):
 | |
|     tests.addTests(loader.loadTestsFromNames(
 | |
|         create_testcases_for_templates()))
 | |
|     return tests
 | |
| 
 | |
| qubes.tests.maybe_create_testcases_on_import(create_testcases_for_templates)
 | 
