Another thought that struck me as far as the (int) constraint is that I can have a number of ArrayLists that are embedded in a an object that has the first and last index in it. These would be linked together to provide unconstrained (MAX_VALUE or 2,147,483,647 lists) storage

8:30 – 4:30 BRI

I realized yesterday that the Ingest and Query microservices need to access the same GeoMesa Spring service. That keeps all the general store/query GeoMesa access code in one place, simplifies testing and allows for DI to provide the correct (hbase, accumulo, etc) implementation through a facade interface.

Got tangled up with getting classpaths right and importing the proper libraries

Got the maven files behaving, or at least not complaining on mvn clean and mvn compile!

Well that’s a new error: Error: Could not create the Java Virtual Machine. I get that running the new installation with the geomesa-quickstart-hbase

Ah, that’s what will happen when you paste your command-line arguments into the VM arguments space just above where it should go…

Wednesday’s goal will to verify that HBaseQuickStart is running correctly in its new home and start to turn it into a service.

One of the most important developments affecting electoral competition in the United States has been the increasingly partisan behavior of the American electorate. Yet more voters than ever claim to be independents. We argue that the explanation for these seemingly contradictory trends is the rise of negative partisanship. Using data from the American National Election Studies, we show that as partisan identities have become more closely aligned with social, cultural and ideological divisions in American society, party supporters including leaning independents have developed increasingly negative feelings about the opposing party and its candidates. This has led to dramatic increases in party loyalty and straight-ticket voting, a steep decline in the advantage of incumbency and growing consistency between the results of presidential elections and the results of House, Senate and even state legislative elections. The rise of negative partisanship has had profound consequences for electoral competition, democratic representation and governance.

Working on putting together an indexable high-dimension matrix that can contain objects. Generally, I’d expect it to be doubles, but I can see Strings and Objects as well.

Starting off by seeing what’s in the newest Apache Commons Math (v 3.6.1)

Working out how to add capability to the sim for P&RCH paper. My thoughts from vacation:

The agents contribution is the heading and speed

The UI is what the agent’s can ‘see’

The IR is what is available to be seen

An additional part might be to add the ability to store data in the space. Then the behavior of the IR (e.g. empty areas) would b more apparent, as would the effects of UI (only certain data is visible, or maybe only nearby data is visible) Data could be a vector field in Hilbert space, and visualized as color.

Updated IntelliJ

Working out how to to have a voxel space for the agents to move through that can also be drawn. It’s any number of dimensions, but it has to project to 2D. In the case of the agents, I just choose the first two axis. Each agent has an array of statements that are assembled into a belief vector. The space can be an array of beliefs. Are these just constructed so that they fill a space according to a set of rules? Then the xDimensionName and yDimensionName axis would go from (0, 1), which would scale to stage size? IR would still be a matter of comparing the space to the agent’s vector. Hmm.

Voters are often highly dependent on partisanship to structure their preferences toward political candidates and policy proposals. What conditions enable partisan cues to “dominate” public opinion? Here I theorize that variation in voters’ reliance on partisanship results, in part, from the opportunities their environment provides to learn about politics. A conjoint experiment and an observational study of voting in congressional elections both support the expectation that more detailed information environments reduce the role of partisanship in candidate choice.

9:00 – 5:00 BRI

Good lord, the BoA corporate card comes with SIX seperate documents to read.

Onward to Chapter Three and Spring database interaction

Well that’s pretty clean. I do like the JdbcTemplate behaviors. Not sure I like the way you specify the values passed to the query, but I can’t think of anything better if you have more than one argument:

Created a timer callback, modified the dataprovider and re-rendered the chart. Seems to be working fine, with no flicker, even with a 100Msec update rate.

11:00 – 1:00 FP

Meeting with Andrea. Looks like uniqueness can be determined by word choice in a corpus as small as 500. That does make things easier. It also allows for triangulation against other metrics, which would allow for looking at accelerometer data from several body positions for example. Though I’m not sure that’s needed.

An issue to consider is that people who might use the site would have other examples of their writing. This means that an anonymous source could be identified. As a way around this, a vector-based translation algorithm could be trained using identification code to remove/modify the parts of the user’s language that are identifiable.

And actually, this means that I could build a simple website that you train once, then write to. The site then transpiles the user’s words and publishes to twitter or Facebook for example. This just addresses the anonymous part of the problem, not the trustworthiness part, but it’s nice low-hanging fruit.

1:00 – 4:00

Javascript

Added a datatable, which behaves better. For example, the set() method fires a refresh. Spent the rest of the day trying to get the same behavior with the chart and finally wound up posting on the forums.

Cognos data is wrong in that amounts in excess of the budget are being calculated from expenditures. The >100% rule might fix this, but I’m worried that in may need to be more sophisticated.

Did a Cognos upload of 1300 records with logging turned on that really slowed down Tomcat, including things like logging in. Accessing the published data REST servlet on the server worked fine though. Not sure what the deal is. I turned off logging to the DB and restarted the server, which sped things up. We need to try another Cognos ingest to see what happens. Lenny did one this morning that took a fraction of the time of the later one.

At the very least, we’re going to need to deal with user reactions to an apparently hung system with a dialog or something.

In meeting with Dr. Kuber, I brought up something that I’ve been thinking about since the weekend. The interface works, provably so. The pilot study shows that it can be used for (a) training and (b) “useful” work. If the goal is to produce “blue collar telecommuting”, then the question becomes, how do we actually achieve that? A dumb master-slave system makes very little sense for a few reasons:

Time lag. It may not be possible to always get a fast enough response loop to make haptics work well

Machine intelligence. With robots coming online like Baxter, there is certainly some level of autonomy that the on-site robot can perform. So, what’s a good human-robot synergy?

I’m thinking that a hybrid virtual/physical interface might be interesting.

The robotic workcell is constantly scanned and digitized by cameras. The data is then turned into models of the items that the robot is to work with.

These items are rendered locally to the operator, who manipulates the virtual objects using tight-loop haptics, 3D graphics, etc. Since (often?) the space is well known, the objects can be rendered from a library of CAD-correct parts.

The operator manipulates the virtual objects. The robot follows the “path” laid down by the operator. The position and behavior of the actual robot is represented in some way (ghost image, warning bar, etc). This is known as Mediated Teleoperation, and described nicely in this paper.

The novel part, at least as far as I can determine at this point is using mediated telepresence to train a robot in a task:

The operator can instruct the robot to learn some or all of a particular procedure. This probably entails setting entry, exit, and error conditions for tasks, which the operator is able to create on the local workstation.

It is reasonable to expect that in many cases, this sort of work will be a mix of manual control and automated behavior. For example, placing of a part may be manual, but screwing a bolt into place to a particular torque could be entirely automatic. If a robot’s behavior is made fully autonomous, the operator needs simply to monitor the system for errors or non-optimal behavior. At that point, the operator could engage another robot and repeat the above process.

User interfaces that inform the operator when the robot is coming out of autonomous modes in a seamless way need to be explored.