Does RTC source control support statistics reporting? - rtc

Does Rational Team Concert source control support statistics reporting like svnstats?

Related

What‘s the difference between OSI and OPC?

OSI: Open Systems Interconnection Reference Model
OPC: OLE (Object Linking and Embedding) Process Control
I can‘t figure out the difference, since both of these refer to communication between machines.
Glad about any answer. Thanks
The Open Systems Interconnection (OSI) Model is a logical layout to define how systems should communicate with each other, using various protocols available at each of seven layers. It is a universal language to guide vendors and IT professionals to develop products and solutions or troubleshoot problems in networks. So, it is more like a conceptual tool than a specific protocol specification or a single technology.
On the other hand, OPC (Open Platform Communications, formerly OLE Process Control) can be described as an industrial standard M2M protocol for interoperability. The OPC standard is a series of specifications with a long history. The platform-dependent old specifications are called OPC Classic, and the next generation of OPC technology is known as OPC UA. In other words, it defines a specific communication solution in order to address a certain gap in industry.
Consequently, as a reference model and a generic tool, we can use the OSI model to develop a better understanding of how OPC deals with the problems we have in industrial networks today.

Low performance of the new Graphical Views in Dynamic Workload Scheduler

have you got any tips to optimize the performance of the new Graphical Views when managing hundreds of objects?
If you work with job streams with more than 500 objects (jobs, job streams and dependencies), to optimize the application performance of the new Graphical Views it is recommended that you follow some tips:
Choose Google Chrome
Chrome has better performance compared to other browsers. In some cases, it is 4 times faster than other browsers to open Graphical Views containing complex objects. Also, using Mozilla Firefox is preferable to Microsoft Internet Explorer or Microsoft Edge.
Customize and save the job stream layout
When viewing job streams in the Job Stream View, first customize and save the job stream layout in the Workload Designer modelling graphical view and then open it in the Job Stream View where the same layout is reused to achieve better performance.
Use Graphical Views on a powerful client in terms of CPU and RAM
Adequately set the hardware configuration of the client machine from which the Dynamic Workload Console is accessed.

Tool for testing usability design of a system architecture

I have this payroll system architecture that uses three tier architecture. Our instructor requires as to look for a tool that can test the usability of the design of the system architecture of the payroll system. I am having problem looking for this tool. Can anyone suggests what tools can be used for measuring usability of the payroll system architecture.
Go to webusability.com, then click on usability tools, and then click on card sorting and navigation. Click on either Treejack or Tree Testing. Both of these tools will allow you to measure the goodness of the system architecture from the user's point of view.

Is IBM Rational different from IBM Rational Rose?

As per my knowledge Rational is a testing tool and Rational Rose is a modelling tool.
Rational is a company that makes tools for software development life cycle. Rational Rose is one such tool built by Rational.
Rational is now part of IBM. See www.rational.com.
Rational Software (a company that was acquired by IBM about 7 years ago) and remains a brand in the IBM Software Group. The Rational portfolio is quite broad concentrating on overall Application Lifecycle Management and System Management solutions. Testing, Quality Management, Configuration Management, Mobile Development, Requirements Management, Change Management, Agile Development, Mainframe Development, Enterprise Modernization and much more.
There is IBM Rational Test Real Time , and it is used for module and unit tests , as well as performance and memory profiling . it is mainly a testing tool . know nothing about rational Rose .

Official Kinect SDK vs. Open-source alternatives

Where do they differ?
What are the advantages of choosing libfreenect or OpenNI+SensorKinect, for example, over the Official SDK, and vice-versa?
What are the disadvantages?
Please note that the below answer is per date and some facts may very well be outdated in the near future. Current state of the Official Kinect SDK is beta 1.00.12.
The first obvious difference is that the official SDK is maintained by the Microsoft Research team while OpenKinect is an open source SDK maintained by the open source community. Both has its cons and pros.
The Official SDK is developed by Microsoft which also develops the hardware and therefore should know internal information about the device that the open source society must reverse engineer. Obviously this is to Microsoft's advantage.
Microsoft is pouring a lot of money into this device and I am sure that they will do what they feel is necessary to keep their SDK up to par. Having economy behind it gives many advantages.
On the other hand, never underestimate the force of the open source society: "The OpenKinect community consists of over 2000 members contributing their time and code to the Project. Our members have joined this Project with the mission of creating the best possible suite of applications for the Kinect. OpenKinect is a true "open source" community!" - http://openkinect.org/wiki/Main_Page.
OpenKinect was released long before the official SDK as the kinect device was hacked on the first or second day of its release. Kudos to OpenKinect!
Programming languages supported:
Official SDK: C++, C#, or Visual Basic by using Microsoft Visual Studio 2010.
OpenKinect: Python, C, C++, C#, Java, Lisp and more! Obviously not requiring Visual Studio.
Operating systems support:
Official SDK: only installs on Windows 7.
OpenKinect: runs on Linux, OS X and Windows
Clearly advantage OpenKinect.
License:
The Official SDK is in its current beta state only for testing. The SDK has been developed specifically to encourage wide exploration and experimentation by academic, research and enthusiast communities. commercial applications are not permitted. Note however that this will probably change in future releases of the SDK. Visit the FAQ for more information
OpenKinect appers to be open for commercial usage, but online sources state that it may not be that simple. I would take a good look at the terms before releasing any commercial apps with it. Read Kinect – Licensing implications of open hardware projects for more info.
Documentation and support:
Official SDK: well documented and provides a support forum
OpenKinect: appears to have a mailing list, twitter and irc. but no official forum/QA? Documentation on website is not as rich as I would like it to be.
Device calibration:
Different Kinect devices may differ slightly depending on the batch that they were produced in. Thus device calibration is sometimes required. But:
the Official SDK does not provide any calibration settings but I have so far not had to calibrate the device I am working on. According to something I read online (link lost) at production time the calibration parameters are written to the kinect device, so with the Official SDK calibration is not needed.
OpenKinect features device calibration: http://openkinect.org/wiki/Calibration. Thus I believe that you should calibrate your device if you go with OpenKinect.
If its true that calibration is only needed for OpenKinect that is a big advantage for the official SDK as it is easier to distribute and install applications without such need.
Personally, after a failed try with the OpenKinect SDK I went with the official SDK, which
came with drivers that installed out of the box
came with examples and code for easy getting into business
All-in-all: I could start my own development within 15 minutes or so.
Now, after working with the Kinect for a few months, I have to say that I am quite satisfied with the API provided. I cannot however compare it to the OpenKinect SDK as I in fact never got it working (but perhaps it didn't give it a fair try).
UPDATE: As of February 1st 2012 there is a commercial license for the official SDK:
"The commercial license for this release authorizes development and distribution of commercial applications. The prior SDK was a beta, and as a result was appropriate only for research, testing and experimentation, and was not suitable for use with a final, commercial product. The new license will enable developers to create and sell their Kinect for Windows applications to end user customers using Kinect for Windows hardware on Windows platforms."
Developer Frequently Asked Questions
As explained by Avada Kedavra in his/her answer, these are some interesting differences:
supported operating systems: you can only use Microsoft SDK on Windows, while open source solutions are usually able to work on other operating systems;
programming languages: you have a wider choice with open source solutions, while Microsoft only supports C++ and C# (Visual Basic is no more supported with SDK 2.0);
documentation and support: Microsoft offer a good forum and a well done documentation (with a lot of samples); but there are several open source solution well documented;
license: Microsoft is less or more proprietary, open source is less or more free. Consider also that open source ideas have sometimes been bought by big companies, and transformed in something that is no more open. Probably yours will not be the case, but keep in mind this additional eventuality.
In my personal opinion, the most significant difference between open source solutions and Microsoft SDKs is strictly related to the skeletal tracking algorithm.
While depth and RGB data can be effectively provided by both open/free APIs and Microsoft SDKs, implementing skeletal tracking capabilities is not only a matter of reverse engineering.
To implement such an algorithm, developers must have strong competences in pattern recognition and machine learning areas, and I am quite sure that such kind of knowledge is available among the open source community. But the implementation of skeletal tracking is based on a "trained" algorithm, that requires a lot of experiments to collect very large amount of data. These data are then used to "train" the algorithm, that can recognize the skeletal joints.
Getting enough data, but also adjusting and properly using them, requires a lot of time and money. Microsoft researchers and developers are in the best conditions to work on this kind of stuff, simply because it is their job.
In my previous experiences, I noticed that open source solutions provide good skeletal tracking capabilities, but they are not at the same level of what Microsoft offers with its SDK.
Remember also that Microsoft SDK provide a lot of additional capabilities, like facial recognition or joint orientation, and several widgets very useful if you want to fastly build a gestural GUI.
So what I suggest is: if you are working on a project in which you simply need depth and/or RGB data, or if you have the necessity to use a programming language that is not supported by Microsoft SDK, then you should opt for open source solution. Otherwise, Microsoft SDK would be my best choice.
I would strongly recommend the Cinder framework. (libcinder.org)
It supports both OpenNI and Kinect develoment, if you're using C++. It now supports Kinect SDK 1.7 and OpenNI 2, via these Cinderblocks:
MS Kinect SDK 1.7 (stable)
https://github.com/BanTheRewind/Cinder-MsKinect
OpenNI 2 / NITE 2.2 (alpha)
https://github.com/wieden-kennedy/Cinder-OpenNI
Both can do skeletal tracking out of the boz, OpenNI being capable of tracking up to six skeletons simultaneously. OpenNI 2 is gaining rapidly on the Kinect, although the new Kinect will probably change that when it comes out next month. However the basic underlying principles are unlikely to change.
The main drawback with the initial release of OpenNI was that it required a full body activation pose to recognise a user, which was a deal breaker for a lot of applications - however this seems to have been solved in the newer versions and OpenNI 2 also supports robust hand tracking at close range, although it still requires a focus gesture initially. If you work on Mac or Linux, it's pretty much your only choice.