Search Results

Search found 32567 results on 1303 pages for 'test manager'.

Page 271/1303 | < Previous Page | 267 268 269 270 271 272 273 274 275 276 277 278  | Next Page >

  • Javascript Module pattern with DOM ready

    - by dego89
    I am writing a JS Module pattern to test out code and help me understand the pattern, using a JS Fiddle. What I can't figure out is why my "private methods" on line 25 and 26, when referenced via DOM ready, have a value of undefined. JSFiddle Code Sample: var obj = { key: "value" }; var Module = (function () { var innerVar = "5"; console.log("obj var in Module:"); console.log(obj); function privateFunction() { console.log("privateFunction() called."); innerFunction(); function innerFunction() { console.log("inner function of (private function) called."); } } function _numTwo() { console.log("_numTwo() function called."); } return { test: privateFunction, numTwo: _numTwo } }(obj)); $(document).ready(function () { console.log("$ Dom Ready"); console.log("Module in Dom Ready: "); console.log(Module.test()); });

    Read the article

  • How to automatically mount a folder and change ownership from root in virtualbox

    - by Fiztban
    It is my first time using virtualbox and ubuntu (14.04), I am on a host Windows 7 OS. I am trying to mount a shared folder that has files I need to access both in the virtualbox and on the windows OS. I have successfully mounted them using the vboxsf from the Guest Additions installed. To mount I used the command sudo mount -t vboxsf <dir name in vbox> <directory in linux for example I used sudo mount -t vboxsf Test /home/user/Test I found several ways of mounting the directories automatically upon startup using for example the /etc/rc.local method (here) where you modify said file appending the command to it (without sudo). Or by using the fstab method (here). I prefer the rc.local method personally. Once mounted it has permissions dr-xr-xr-x however once mounted the directory is of root ownership and chown user /home/user/Test has no effect. This means I cannot make or change files in it as a normal user. In the VirtualBox the directory to be shared is not set as read-only. Is there a way to automatically mount the shared folder and assign ownership to my non root user?

    Read the article

  • upower wants to uninstall lots of packages

    - by Phoeey
    I'm running Natty and have a problem with upower. Currently installed is upower 0.9.5-4 (Maverick?), with upower 0.9.9-4 available. Update Manager won't upgrade it, but if I go to Synaptic and use Ctrl+E (force version) it prompts to remove all of the following packages: gdm gdm-guest-session gnome-power-manager gnome-session gnome-session-bin indicator-session nautilus-share ubuntu-desktop This seems like a fair chunk of the GUI, so I'm not keen to let it go ahead. Is there a better/safer way to force the package upgrade? NOTE: This machine was upgraded from Maverick to Natty using the Alternate CD a while ago. It locked-up about 80-90% through the upgrade requiring the machine to be hard reset, but it finished the upgrade process after rebooting. I was making sure everything was OK before going to Oneiric when I discovered this little gem.

    Read the article

  • Why doesn't this batch file work for me?

    - by asdfg
    The following batch file is not woking. @echo off python -c "print('echo text')" %TEMP%\test.bat call %TEMP%\test.bat Can anyone help me with this?. edit: I needed unix eval functionality in windows. I could not find a direct way though. So I redirected the eval string to a temporary batch file and executed it. The temporary batch file was successfully created but calling it did not work in the above case. I noticed that any command after the test.bat creation did not work.

    Read the article

  • Hadoop:Only master node does the work

    - by user287722
    I've setup a Hadoop 2.2 cluster with 1 master node(namenode and secondary namenode) and 3 slave nodes(datanode and namenode on each one).All of the machines use Linux Mint 64bit. When I run my MapReduce program, writen in Java, I can only see that master node is using extra CPU and RAM. Slave nodes are not doing a thing. I've checked the logs from all of the namenodes and there is nothing wrong with the namenodes on slave nodes. Resource Manager is running and all of the slave nodes can see the Resource Manager. I used this http://n0where.net/hadoop-2-2-multi-node-cluster-setup/ tutorial to configure my nodes. Datanodes are working in terms of distributed data storing but I can't see any indication of distributed data processing. Do I have to configure the xml configuration files in some other way so all of the machines will process data while I'm running my MapReduce Job?

    Read the article

  • Unity Launcher missing and Terminal shortcut stopped working

    - by Diana
    When I restarted my Ubuntu after it being updated, I found the Unity launch bar was gone. I tried looking up several solutions online, blindly typing in the commands that others suggested, and I found after the umpteenth time rebooting after typing in another foreign command, I found I couldn't even open up the terminal with the Ctrl+Alt+t shortcut either and I have no idea how to fix this. This was the last set of instructions I remember following before terminal stopped opening. Install CCSM sudo apt-get install compizconfig-settings-manager And then run it in terminal (Ctrl+Alt+t) ccsm Then search for "Ubuntu Unity Plugin". If it's not checked, check it. If it is checked you may have another issue. Try restarting compiz: kill -9pidof compiz` This will kill it, it should start up again, if it doesn't, CTRL+ALT+F2 and then type: sudo service lightdm restart to restart the window manager. Go back to the terminal and run: unity --reset

    Read the article

  • Force apt-get to ask all install equations again?

    - by user204744
    I made a mistake. I answered the wrong way when installing something with apt-get. Now, when I purge it and reinstall it, it doesn't ask me the question any more. How do I force apt-get to ask me the question again? I'd really rather do this through the package manager than making the changes manually; that's the whole point of using a package manager.... The package in question is jackd. The first time I installed it, I answered "no" to the question about real-time priority. Now, I wish I'd said yes. However, even though I purge and install it again, I don't get the question. (I have tried dpkg-reconfigure jackd -- that doesn't give me any questions to answer.)

    Read the article

  • Is it dangerous for me to give some of my Model classes Control-like methods?

    - by Pureferret
    In my personal project I have tried to stick to MVC, but I've also been made aware that sticking to MVC too tightly can be a bad thing as it makes writing awkward and forces the flow of the program in odd ways (i.e. some simple functions can be performed by something that normally wouldn't, and avoid MVC related overheads). So I'm beginning to feel justified in this compromise: I have some 'manager programs' that 'own' data and have some way to manipulate it, as such I think they'd count as both part of the model, and part of the control, and to me this feels more natural than keepingthem separate. For instance: One of my Managers is the PlayerCharacterManager that has these methods: void buySkill(PlayerCharacter playerCharacter, Skill skill); void changeName(); void changeRole(); void restatCharacter(); void addCharacterToGame(); void createNewCharacter(); PlayerCharacter getPlayerCharacter(); List<PlayerCharacter> getPlayersCharacter(Player player); List<PlayerCharacter> getAllCharacters(); I hope the mothod names are transparent enough that they don't all need explaining. I've called it a manager because it will help manage all of the PlayerCharacter 'model' objects the code creates, and create and keep a map of these. I may also get it to store other information in the future. I plan to have another two similar classes for this sort of control, but I will orchestrate when and how this happens, and what to do with the returned data via a pure controller class. This splitting up control between informed managers and the controller, as opposed to operating just through a controller seems like it will simplify my code and make it flow more. My question is, is this a dangerous choice, in terms of making the code harder to follow/test/fix? Is this somethign established as good or bad or neutral? I oculdn't find anything similar except the idea of Actors but that's not quite why I'm trying to do. Edit: Perhaps an example is needed; I'm using the Controller to update the view and access the data, so when I click the 'Add new character to a player button' it'll call methods in the controller that then go and tell the PlayerCharacterManager class to create a new character instance, it'll call the PlayerManager class to add that new character to the player-character map, and then it'll add this information to the database, and tell the view to update any GUIs effected. That is the sort of 'control sequence' I'm hoping to create with these manager classes.

    Read the article

  • After Upgrade from 13.04 to 13.10 missing user in login screen

    - by Mark
    I have upgraded to 13.10 from 13.04 now after booting (even after updates and rebooting) my user id is not listed. It has a uid higher than the minimum in lightdm.conf. I have the user light display manager , guest , and remote login options in the login screen but not my ususal user id that I have been using from ubuntu 9 something. It is a normal user and an administrator so that I could authorize drivers, updates, etc... I do not have any custom display manager setup. It has been standard unity since unity was available. I can login to my account by logging in as guest opening a termial and using su - user Any insight would be appreciated. Thanks

    Read the article

  • Pillar Axiom OpenWorld Highlights

    - by uwes
    During the Storage General Session at Oracle OpenWorld Conference 2012 in San Francisco, the following Axiom-related announcements were made: Oracle Platinum Services for Axiom 600: Extending Oracle's Platinum Services to Axiom 600 as a standalone product –  the same level of service and support you get with Exadata – 24/7 fault monitoring, dedicated response and escalation management to meet enterprise-grade SLA’s, patch planning and management. Oracle Enterprise Manager Axiom Plug-in: Allowing DBA's to manage, maintain, monitor and provision the Axiom 600 storage system from Oracle EM. Oracle Virtual Machine Axiom Plug-in: Allowing Oracle VM and System Administrators to manage, maintain, monitor and provision the Axiom 600 storage system from Oracle VM using Storage Connect. Oracle Axiom Data Protection Manager 3.1: Leveraging Axiom's Copy Services, System Administrators can automatically create Application Consistent Clones of critical Windows and Oracle DataBase environments for quick recovery. For More Information Go To: Oracle.com Pillar Axiom Page Oracle Technology Network SAN Page

    Read the article

  • Script to run chown on all folders and setting the owner as the folder name minus the trailing /

    - by Shikoki
    Some numpty ran chown -R username. in the /home folder on our webserver thinking he was in the desired folder. Needless to say the server is throwing a lot of wobbelys. We have over 200 websites and I don't want to chown them all individually so I'm trying to make a script that will change the owner of all the folders to the folder name, without the trailing /. This is all I have so far, once I can remove the / it will be fine, but I'd also like to check if the file contains a . in it, and if it doesn't then run the command, otherwise go to the next one. #!/bin/bash for f in * do test=$f; #manipluate the test variable chown -R $test $f done Any help would be great! Thanks in advance!

    Read the article

  • Changing Wallpaper on 12.04 Login Screen

    - by userIsAMonkey
    I'm using this link but seems to be not working on 12.04, here's the Terminal message below: Are there other softwares/tips for changing the login screen? I'm also using link but seems outdated. Failed to fetch http://ppa.launchpad.net/claudiocn/slm/ubuntu/dists/precise/main/source/Sources 404 Not Found W: Failed to fetch http://ppa.launchpad.net/claudiocn/slm/ubuntu/dists/precise/main/binary-amd64/Packages 404 Not Found W: Failed to fetch http://ppa.launchpad.net/claudiocn/slm/ubuntu/dists/precise/main/binary-i386/Packages 404 Not Found E: Some index files failed to download. They have been ignored, or old ones used instead. law@ubuntu:~$ sudo apt-get install simple-lightdm-manager Reading package lists... Done Building dependency tree Reading state information... Done E: Unable to locate package simple-lightdm-manager

    Read the article

  • Setting Default Ubuntu eth port

    - by No1_Melman
    Background: I have a desktop computer running Ubuntu 12.04. I have installed DRBL to handle DHCP server and Clonezilla Server. I have a network card in my desktop with 2 ports in it. These 2 ports are bonded together and are used by DRBL. There is a 3rd port integrated in the motherboard. I want to use this 3rd port for internet. DRBL knows that this is set for internet. Ubuntu Network-Manager has stopped recognising the connections (device not managed). It use to say that eth1 was default (first port in the network card) I want eth0 to be default as that is the Internet port. Question: How do I make network-manager see eth0 as the defualt port?

    Read the article

  • Connected to Access Point but no internet connection?

    - by ReaZem
    I am trying to use my school's WPA Enterprise network to access the internet with my Three-OS laptop. I currently have Windows7, Ubuntu11.10 and Backtrack5 installed. I can use my home's WPA2 network without any issues with the 3 OS. I can successfully connect to my school's network with the 3 OS but I can only browse the internet with windows7. Ubuntu network-manager and Backtrack WICD both say that I am connected to the Acces Point but I cannot browse the internet. I have tried uninstalling network-manager and installing WICD but it did not solve the problem. Every time I connect to the Access Point, I am able to browse the internet for about 10 seconds. I have tried disabling IPV6 I have tried changing DNS.

    Read the article

  • How to get Apache to follow symlink instead of downloading it?

    - by user792445
    I am just using the standard apache config file which mentions that it follows symlinks, but when I hit the url http://localhost/test it downloads the symlink file instead of following it. What config do I need to change to get apache to follow the symlink instead of downloading it? This is an ls on the directory: $ ls -al total 10 drwx------+ 1 SYSTEM SYSTEM 0 Oct 20 10:55 . drwx------+ 1 SYSTEM SYSTEM 0 Aug 26 12:27 .. -rw-r--r--+ 1 me None 47 Oct 20 10:14 index.html lrwxrwxrwx 1 me None 29 Oct 19 17:10 test -> /home/me/projects/test This is in my apache config file: <Directory "D:/Program Files (x86)/Apache Software Foundation/Apache2.2/htdocs"> Options Indexes FollowSymLinks AllowOverride None Order allow,deny Allow from all </Directory>

    Read the article

  • "An upgrade from 'precise' to 'oneiric' is not unsupported with this tool" error while upgrading

    - by tas
    A brief electrical outage interrupted my 12.04 upgrade. After many tries, which included trying to upgrade from a command line, I was able to access Ubuntu screens again. I find that the 12.04 upgrade was indeed not completed; I am still running 11.10. However, Update Manager does not give me the option of installing 12.o4 now. And, when I run Update Manager, I am asked if I want to execute a "partial install." When I answer "yes," I receive the message above. Besides a clean install, any ideas about what can I do to upgrade to 12.4?

    Read the article

  • Blank desktop after updates today, only unity2d works now [closed]

    - by NewUbuntuUser
    Possible Duplicate: Unity doesn't load, no Launcher, no Dash appears I have been using 12.04 (wubi) since a week now and this is my 1st exposure to linux. Everything was going on fine till now , but today as soon as I updated through the update manager , docky gave a message that compositing is required or something, and i got an error window which asked me to report the error and then the update manager asked me to reboot. After rebooting , i just get a blank desktop screen , no launcher, no toolbar. I have to restart with the power key. However if i login through Unity 2d everything is fine except the benefits of the 3d environment. I guess something got messed up after the update and i cant figure out which program or file caused this mess. I would highly appreciate if someone could help me out with this as I really liked working on ubuntu after windows 7. Thanks! SOLVED-- Thanks @jrg for the link provided.. it helped me to come out of this mess. Actually some update of compizconfig made the unity plugin inactive did something to the Animation add ons , the one which you use for the burn effect etc. What i did was : On the blank desktop Pressed keys Ctrl + Alt + T to bring up the terminal and typed ccsm This brought up the compiz config system manager, there i enabled the unity plugin and rebooted. Everything started working fine. But then again when i started the addons plugin ,everything went back to square one.. :( . This time pressing Ctrl + Alt + T also did not help. Then i tried Ctrl + Alt + F1 , it brought up a terminal or whatever u call it, and then i typed unity --reset , it did some resetting and in the end it showed some comositing done , I pressed Ctrl + Alt + F7 to come back to the desktop , and here it was , my old sweet desktop.. it had the animation effects too like wobbly windows, excepting the addons like the burn effect, i guess something got wrong with the last update of that addon plugin and now whenever i try turning that on , everything goes poof!! 6 hours wasted , but i guess i learnt something new.. not bad for an orthodontist i guess :))

    Read the article

  • Is there a better way to organize my module tests that avoids an explosion of new source files?

    - by luser droog
    I've got a neat (so I thought) way of having each of my modules produce a unit-test executable if compiled with the -DTESTMODULE flag. This flag guards a main() function that can access all static data and functions in the module, without #including a C file. From the README: -- Modules -- The various modules were written and tested separately before being coupled together to achieve the necessary basic functionality. Each module retains its unit-test, its main() function, guarded by #ifdef TESTMODULE. `make test` will compile and execute all the unit tests, producing copious output, but importantly exitting with an appropriate success or failure code, so the `make test` command will fail if any of the tests fail. Module TOC __________ test obj src header structures CONSTANTS ---- --- --- --- -------------------- m m.o m.c m.h mfile mtab TABSZ s s.o s.c s.h stack STACKSEGSZ v v.o v.c v.h saverec_ f.o f.c f.h file ob ob.o ob.c ob.h object ar ar.o ar.c ar.h array st st.o st.c st.h string di di.o di.c di.h dichead dictionary nm nm.o nm.c nm.h name gc gc.o gc.c gc.h garbage collector itp itp.c itp.h context osunix.o osunix.c osunix.h unix-dependent functions It's compile by a tricky bit of makefile, m:m.c ob.h ob.o err.o $(CORE) itp.o $(OP) cc $(CFLAGS) -DTESTMODULE $(LDLIBS) -o $@ $< err.o ob.o s.o ar.o st.o v.o di.o gc.o nm.o itp.o $(OP) f.o where the module is compiled with its own C file plus every other object file except itself. But it's creating difficulties for the kindly programmer who offered to write the Autotools files for me. So the obvious way to make it "less weird" would be to bust-out all the main functions into separate source files. But, but ... Do I gotta?

    Read the article

  • A Single wireless network is not showing up in Ubuntu that used too, and other devices have no issue finding or connecting to it

    - by Kit Sunde
    Up until about a week ago I could connect to my wireless, but now I suddenly can't. Every other wireless in the area show up in the network manager. My kindle can connect to it, my friend can connect to it (in windows) without having connected to it before. My other netbook (Samsung NC10) can connect to it, but my other Samsung N150 now cannot even see it. I have tried deleting the old auto entry for it in the network manager Tried disabling and enabling both the networking and the wireless through the network indicator. Disabled the networking through the hardware shortkut key thing [fn] + [F9] Attempting to connect to it as if it was a hidden network. I'm at a loss, how would I go about debugging and unborking my computer?

    Read the article

  • Windows not closing

    - by bhanu
    After login, a number of windows such as Desktop, Downloads, and Pictures opens automatically. When I try to close them one by one closes itself but last one opens all indicating starting file manager in lower panel/status bar. After login all the windows such as desktop,download,home opens automatically when i try to minimise them all minimise. when i try to close them one after another they disappears but not last one which closes itself and reappear along with other indicating message starting file manager in status bar. ( it seems that there are a number of open nautilus windows at startup upon login, and when the user tries to close them, nautilus crashes and tries to restart)

    Read the article

  • 13.10 Desktop - Stop Suspension on Lid Closure

    - by bpmitche
    I'm having a problem with power management settings on my ASUS K52-F Laptop. Upgraded from 13.04 to 13.10 earlier today; closing the laptop lid puts the laptop into suspension (even on AC Power - tried fixing it in gnome power settings, tried fixing it in Cinnamon power settings, tried changing it in DCONF-Editor, but no matter where I change the setting the laptop lid still suspends). After resuming from suspend, I restore to the gnome window/login manager (instead of Cinnamon login manager). Wifi doesn't reconnect, but the icon in Cinnamon shows that I'm connected. IFCONFIG shows that the interface is up and working normally, but still not reconnecting. ifconfig wlan0 down / up doesn't change anything. Any help would be greatly appreciated.

    Read the article

  • How do I upgrade from 9.04 to 10.04.2?

    - by Yadnesh
    I'm currently runing Ubuntu 9.04 Jaunty. I want to upgrade to 10.04.02, but whenever I use the Update manager to do this, it fails with the error "An upgrade from 'jaunty' to 'lucid' is not supported with this tool". I also tried to run sudo do-release-upgrade -d, but it fails with the same error message: Checking for a new ubuntu release Done Upgrade tool signature Done Upgrade tool Done downloading extracting 'lucid.tar.gz' authenticate 'lucid.tar.gz' against 'lucid.tar.gz.gpg' tar: Removing leading `/' from member names Reading cache Checking package manager Can not upgrade An upgrade from 'jaunty' to 'lucid' is not supported with this tool.

    Read the article

  • Is it OK to introduce methods that are used only during unit tests?

    - by Mchl
    Recently I was TDDing a factory method. The method was to create either a plain object, or an object wrapped in a decorator. The decorated object could be of one of several types all extending StrategyClass. In my test I wanted to check, if the class of returned object is as expected. That's easy when plain object os returned, but what to do when it's wrapped within a decorator? I code in PHP so I could use ext/Reflection to find out a class of wrapped object, but it seemed to me to be overcomplicating things, and somewhat agains rules of TDD. Instead I decided to introduce getClassName() that would return object's class name when called from StrategyClass. When called from the decorator however, it would return the value returned by the same method in decorated object. Some code to make it more clear: interface StrategyInterface { public function getClassName(); } abstract class StrategyClass implements StrategyInterface { public function getClassName() { return \get_class($this); } } abstract class StrategyDecorator implements StrategyInterface { private $decorated; public function __construct(StrategyClass $decorated) { $this->decorated = $decorated; } public function getClassName() { return $this->decorated->getClassName(); } } And a PHPUnit test /** * @dataProvider providerForTestGetStrategy * @param array $arguments * @param string $expected */ public function testGetStrategy($arguments, $expected) { $this->assertEquals( __NAMESPACE__.'\\'.$expected, $this->object->getStrategy($arguments)->getClassName() ) } //below there's another test to check if proper decorator is being used My point here is: is it OK to introduce such methods, that have no other use than to make unit tests easier? Somehow it doesn't feel right to me.

    Read the article

  • Drive still usable if Seatools reports errors?

    - by Rob
    I have a Seagate 3TB Expansion Desktop drive that was part of a Linux RAID 6 array that failed. I eventually did a zero fill both through Seagate DiscWizard and via Linux dd, neither reported errors. When I ran Seatools now, I got: Short DST - Started 5/31/2014 10:04:36 PM Short DST - Pass 5/31/2014 10:05:37 PM Long Generic - Started 5/31/2014 10:15:19 PM Bad LBA: 518242762 Not Repaired (whole bunch of bad LBAs ommited) Bad LBA: 518715255 Not Repaired Long Generic Aborted 6/1/2014 3:12:18 AM i.e. the short test passed, the long test failed. Unfortunately, the drive is out of warranty, so I can't just RMA it. But I hate tossing a drive that can still be used. So, my questions are: If the zero fill succeeded, and the short test passed, can I still use the whole drive? if not, since I'm using LVM on top of RAID, is there a way to tell either of these to just skip the bad area? If not the above, can I just create partitions before and after the part of the drive with the bad LBAs?

    Read the article

< Previous Page | 267 268 269 270 271 272 273 274 275 276 277 278  | Next Page >